V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] Output code: V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # AOT ID: ['0_inference'] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from ctypes import c_void_p, c_long V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import torch V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import random V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import os V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import tempfile V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from math import inf, nan V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.hooks import run_intermediate_hooks V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.utils import maybe_profile V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.codegen.memory_planning import _align as align V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch import device, empty_strided V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.async_compile import AsyncCompile V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.select_algorithm import extern_kernels V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.codegen.multi_kernel import MultiKernelCall V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] aten = torch.ops.aten V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_ops = torch.ops.inductor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] _quantized = torch.ops._quantized V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride = torch._C._dynamo.guards.assert_size_stride V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] empty_strided_cpu = torch._C._dynamo.guards._empty_strided_cpu V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] empty_strided_cuda = torch._C._dynamo.guards._empty_strided_cuda V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] reinterpret_tensor = torch._C._dynamo.guards._reinterpret_tensor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] alloc_from_pool = torch.ops.inductor._alloc_from_pool V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] async_compile = AsyncCompile() V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/kx/ckxvjo3hhiznc4ieahqn6ez5oq2jhe6or6jnoamh6yry3pp64dc5.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [emb_3], Original ATen: [aten.cat] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # emb_3 => cat V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_cat_0 = async_compile.triton('triton_poi_fused_cat_0', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[512], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*i64', 1: '*fp32', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_cat_0', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_cat_0(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 512 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = xindex < xnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 256 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tl.load(in_ptr0 + (0)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tl.broadcast_to(tmp5, [XBLOCK]) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 0, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 >= tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.full([1], 128, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp0 < tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tmp6.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tmp0.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = -9.210340371976184 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tmp8 * tmp9 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = 0.0078125 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = tmp10 * tmp11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tl_math.exp(tmp12) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = tmp7 * tmp13 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = tmp14 * tmp15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp17 = tl_math.sin(tmp16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18 = tl.full(tmp17.shape, 0.0, tmp17.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19 = tl.where(tmp4, tmp17, tmp18) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp20 = tmp0 >= tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp21 = tl.full([1], 256, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp22 = tmp0 < tmp21 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp23 = (-128) + x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp24 = tmp23.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp25 = tmp24 * tmp9 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp26 = tmp25 * tmp11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp27 = tl_math.exp(tmp26) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp28 = tmp7 * tmp27 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp29 = tmp28 * tmp15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp30 = tl_math.cos(tmp29) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp31 = tl.full(tmp30.shape, 0.0, tmp30.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp32 = tl.where(tmp20, tmp30, tmp31) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp33 = tl.where(tmp4, tmp19, tmp32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp33, xmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_heuristics import grid, split_scan_grid, grid_combo_kernels, start_graph, end_graph V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._C import _cuda_getCurrentRawStream as get_raw_stream V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/ay/caykwk6ibs4q2wisgu3nlt2wrwimvicwlfsqqsfonccmq3gwcupp.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [emb_4, sample, to_2], Original ATen: [aten._to_copy, aten.cat, aten.constant_pad_nd, aten.view] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # emb_4 => cat_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # sample => constant_pad_nd, view_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # to_2 => convert_element_type_3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__to_copy_cat_constant_pad_nd_view_1 = async_compile.triton('triton_poi_fused__to_copy_cat_constant_pad_nd_view_1', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[2048], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*fp32', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused__to_copy_cat_constant_pad_nd_view_1', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused__to_copy_cat_constant_pad_nd_view_1(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 2048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 1024 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 1024) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 256, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.full([1], 0, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp0 >= tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tl.full([1], 128, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tmp0 < tmp5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tmp6 & tmp2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tl.load(in_ptr0 + (128 + (256*x1) + x0), tmp7, eviction_policy='evict_last', other=0.0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp0 >= tmp5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tmp9 & tmp2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tl.load(in_ptr0 + ((256*x1) + ((-128) + x0)), tmp10, eviction_policy='evict_last', other=0.0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = tl.where(tmp6, tmp8, tmp11) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tmp12.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = tl.full(tmp13.shape, 0.0, tmp13.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = tl.where(tmp2, tmp13, tmp14) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp15, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/zz/czzffj47sqhn5iimhr4px5p266u5c2wpnfsrtuofxf6hak6q2da6.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [sample, sample_1, sample_2], Original ATen: [aten.add, aten.constant_pad_nd, aten.silu, aten.view] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # sample => add_2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # sample_1 => convert_element_type_4, convert_element_type_5, mul_5, sigmoid V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # sample_2 => constant_pad_nd_1, view_5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_constant_pad_nd_silu_view_2 = async_compile.triton('triton_poi_fused_add_constant_pad_nd_silu_view_2', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[4096], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_add_constant_pad_nd_silu_view_2', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_add_constant_pad_nd_silu_view_2(in_ptr0, in_ptr1, out_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 4096 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 2048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 2048) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 1152, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1)), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tl.load(in_ptr1 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tmp3 + tmp4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tmp5.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tl.sigmoid(tmp6) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tmp6 * tmp7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp8.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tl.full(tmp9.shape, 0.0, tmp9.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tl.where(tmp2, tmp9, tmp10) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp11, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/kz/ckzvxx7soy6esfx6au7rln5h7czwx6pg24olgmfwjfogdbx33xfn.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [latent], Original ATen: [aten.convolution] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # latent => convolution V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_convolution_3 = async_compile.triton('triton_poi_fused_convolution_3', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[8, 16384], tile_hint=TileHint.SQUARE, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 3), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_convolution_3', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_convolution_3(in_ptr0, out_ptr0, ynumel, xnumel, YBLOCK : tl.constexpr, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ynumel = 8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 16384 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] yoffset = tl.program_id(1) * YBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] yindex = yoffset + tl.arange(0, YBLOCK)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ymask = yindex < ynumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK, YBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] y3 = yindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] y0 = yindex % 4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] y1 = (yindex // 4) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_ptr0 + (x2 + (16384*y3)), ymask, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (y0 + (4*x2) + (65536*y1)), tmp0, ymask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/wt/cwt7sxdkktlnm3e4lgnzgzqerira5sdkczyq6upiuxa7c35e5vqu.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [latent], Original ATen: [aten.convolution] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # latent => convolution V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_tem_fused_convolution_4 = async_compile.triton('triton_tem_fused_convolution_4', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.template( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] num_stages=2, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] num_warps=4, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'kernel_name': 'triton_tem_fused_convolution_4', 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_tem_fused_convolution_4(arg_X, arg_W, out_ptr0): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] KERNEL_H : tl.constexpr = 2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] KERNEL_W : tl.constexpr = 2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] STRIDE_H : tl.constexpr = 2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] STRIDE_W : tl.constexpr = 2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] PADDING_H : tl.constexpr = 0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] PADDING_W : tl.constexpr = 0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] GROUPS : tl.constexpr = 1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] UNROLL : tl.constexpr = False V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ALLOW_TF32 : tl.constexpr = True V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] BLOCK_M : tl.constexpr = 64 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] BLOCK_N : tl.constexpr = 256 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] BLOCK_K : tl.constexpr = 16 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] X = arg_X V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] W = arg_W V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Tensor dimensions V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] BATCH = 2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] IN_C = 4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] IN_H = 128 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] IN_W = 128 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] OUT_C = 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] OUT_H = 64 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] OUT_W = 64 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Strides: V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stride_xn = 65536 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stride_xc = 1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stride_xh = 512 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stride_xw = 4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stride_wc_out = 16 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stride_wc_in = 1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stride_wh = 8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stride_ww = 4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] nhw = tl.program_id(0) * BLOCK_M + tl.arange(0, BLOCK_M) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_y_w = nhw % OUT_W V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] nh = nhw // OUT_W V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_y_h = nh % OUT_H V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_n = nh // OUT_H V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_y_c = tl.program_id(1) * BLOCK_N + tl.arange(0, BLOCK_N) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] group = 0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] GROUP_IN_C = IN_C V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] GROUP_OUT_C = OUT_C V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x_base = X + (group * stride_xc * GROUP_IN_C + idx_n * stride_xn)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] w_base = ( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] W + (group * stride_wc_out * GROUP_OUT_C + idx_y_c * stride_wc_out)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] acc = tl.zeros((BLOCK_M, BLOCK_N), dtype=tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Could be simplified, but slightly slower: V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # for i in range(KERNEL_H): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # for j in range(KERNEL_W): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # for k in range(0, GROUP_IN_C, BLOCK_K): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] BLOCK_K_COUNT = (GROUP_IN_C + BLOCK_K - 1) // BLOCK_K V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] for ijk in range(KERNEL_H * KERNEL_W * BLOCK_K_COUNT): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] k = (ijk % BLOCK_K_COUNT) * BLOCK_K V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ij = ijk // BLOCK_K_COUNT V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] i = ij // KERNEL_W V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] j = ij % KERNEL_W V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_x_h = i - PADDING_H + idx_y_h * STRIDE_H V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_x_w = j - PADDING_W + idx_y_w * STRIDE_W V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_x_c = tl.arange(0, BLOCK_K) + k V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x_ptrs = x_base + ( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] (idx_x_h * stride_xh)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] + (idx_x_w * stride_xw)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] + (idx_x_c * stride_xc)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] mask_x = ( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] (idx_n < BATCH)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] & (idx_x_h >= 0)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] & (idx_x_h < IN_H)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] & (idx_x_w >= 0)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] & (idx_x_w < IN_W)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] & (idx_x_c < GROUP_IN_C)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] matrix_x = tl.load(x_ptrs, mask=mask_x, other=0.0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] w_ptrs = w_base + ( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] (idx_x_c * stride_wc_in)[:, None] + (i * stride_wh) + (j * stride_ww) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] mask_w = (idx_x_c[:, None] < GROUP_IN_C) & (idx_y_c[None, :] < GROUP_OUT_C) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] matrix_w = tl.load(w_ptrs, mask=mask_w, other=0.0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] acc += tl.dot(matrix_x, matrix_w, allow_tf32=ALLOW_TF32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] mask = ( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] (idx_n < BATCH)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] & (idx_y_h < OUT_H)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] & (idx_y_w < OUT_W)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] & (idx_y_c < GROUP_OUT_C)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_n = idx_n[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_c = idx_y_c[None, :] + group * GROUP_OUT_C V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_h = idx_y_h[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] idx_w = idx_y_w[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # inductor generates a suffix V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = idx_w + (64*idx_h) + (4096*idx_c) + (4718592*idx_n) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (tl.broadcast_to(idx_c + (1152*idx_w) + (73728*idx_h) + (4718592*idx_n), acc.shape)), acc, mask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import torch._inductor.kernel.conv V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] meta0 = {'KERNEL_H': 2, 'KERNEL_W': 2, 'STRIDE_H': 2, 'STRIDE_W': 2, 'PADDING_H': 0, 'PADDING_W': 0, 'GROUPS': 1, 'UNROLL': False, 'ALLOW_TF32': True, 'BLOCK_M': 64, 'BLOCK_N': 256, 'BLOCK_K': 16} V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/yr/cyr2uidmv2qdwsxvct57mtyzyd3exxi6oyxcoqocjk6j7r2yfqgu.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add, add_2, mul_4, norm_hidden_states, norm_hidden_states_1], Original ATen: [aten.add, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add => add V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add_2 => add_11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # mul_4 => mul_14 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states => add_10, convert_element_type_10, convert_element_type_11, mul_13, rsqrt, sub_1, var_mean V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_1 => add_12 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_mul_native_layer_norm_5 = async_compile.triton('triton_red_fused_add_mul_native_layer_norm_5', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.reduction( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[8192, 2048], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] reduction_hint=ReductionHint.INNER, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: 'i32', 8: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_mul_native_layer_norm_5', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 12, 'num_reduction': 2, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True} V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_red_fused_add_mul_native_layer_norm_5(in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, out_ptr2, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 8192 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rnumel = 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rbase = tl.arange(0, RBLOCK)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x3 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 4096 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_mean = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_m2 = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_weight = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] for roffset in range(0, rnumel, RBLOCK): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rindex = roffset + rbase V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rmask = rindex < rnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] r2 = rindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr1 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr2 + (r2 + (1152*x0)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 + tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp2 + tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tmp4.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tl.broadcast_to(tmp5, [XBLOCK, RBLOCK]) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_mean_next, tmp7_m2_next, tmp7_weight_next = triton_helpers.welford_reduce( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6, tmp7_mean, tmp7_m2, tmp7_weight, roffset == 0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_mean = tl.where(rmask, tmp7_mean_next, tmp7_mean) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_m2 = tl.where(rmask, tmp7_m2_next, tmp7_m2) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_weight = tl.where(rmask, tmp7_weight_next, tmp7_weight) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_tmp, tmp8_tmp, tmp9_tmp = triton_helpers.welford( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7_mean, tmp7_m2, tmp7_weight, 1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tmp7_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tmp8_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp9_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 4096) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] for roffset in range(0, rnumel, RBLOCK): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rindex = roffset + rbase V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rmask = rindex < rnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] r2 = rindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tl.load(in_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tl.load(in_ptr1 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tl.load(in_ptr2 + (r2 + (1152*x0)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp24 = tl.load(in_ptr3 + (1152 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp25 = tl.load(in_ptr4 + (1152 + r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp26 = tl.load(in_ptr5 + (1152 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp32 = tl.load(in_ptr3 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp33 = tl.load(in_ptr4 + (r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp34 = tl.load(in_ptr5 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = tmp10 + tmp11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = tmp12 + tmp13 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = tmp14.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = tmp15 - tmp7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp17 = 1152.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18 = tmp8 / tmp17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19 = 1e-06 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp20 = tmp18 + tmp19 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp21 = libdevice.rsqrt(tmp20) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp22 = tmp16 * tmp21 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp23 = tmp22.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp27 = tmp25 + tmp26 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp28 = tmp24 + tmp27 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp29 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp30 = tmp28 + tmp29 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp31 = tmp23 * tmp30 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp35 = tmp33 + tmp34 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp36 = tmp32 + tmp35 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp37 = tmp31 + tmp36 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr2 + (r2 + (1152*x3)), tmp37, rmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/wv/cwvinpaomxydoj2prwmcgwqsfbn2qzxt3zbdlxdoavfadffftn52.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key, query, value], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # key => constant_pad_nd_6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # query => constant_pad_nd_5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # value => constant_pad_nd_7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6 = async_compile.triton('triton_poi_fused_constant_pad_nd_6', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16777216], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_6', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_6(in_ptr0, out_ptr0, out_ptr1, out_ptr2, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 16777216 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 2048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 2048) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 1152, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1)), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp3, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr1 + (x2), tmp3, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr2 + (x2), tmp3, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/kl/cklcpz72lzisv4b6xh2s5kk47j7r5qxci2awj5uak4xrzefg6b5t.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_4 => _scaled_dot_product_cudnn_attention V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7 = async_compile.triton('triton_poi_fused__scaled_dot_product_cudnn_attention_7', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16777216], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused__scaled_dot_product_cudnn_attention_7', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused__scaled_dot_product_cudnn_attention_7(in_out_ptr0, in_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 9437184 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_out_ptr0 + (x2), None).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr0 + (x0), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 + tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(in_out_ptr0 + (x2), tmp2, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/xj/cxjmrv4cb4oezymi343qkmkzbtmvx3ao2wbnvim4lfzjcqk42bgb.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_7], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_7 => constant_pad_nd_8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8 = async_compile.triton('triton_poi_fused_constant_pad_nd_8', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16777216], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_8', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_8(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 16777216 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 2048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 2048) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 1152, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + ((72*(x1 % 4096)) + (294912*((x0 // 72) % 16)) + (4718592*(x1 // 4096)) + (x0 % 72)), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp3, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/iz/cizfa5uobinya2cxdo5uzfvrfn6sqrfbvszojkwxvsjr5lf67qcq.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add, attn_output, hidden_states_10, hidden_states_9], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add => add V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # attn_output => mul_15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_10 => add_17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_9 => div_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_9 = async_compile.triton('triton_poi_fused_add_div_mul_9', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16777216], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: '*bf16', 8: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_add_div_mul_9', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 8, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_add_div_mul_9(in_out_ptr0, in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, in_ptr6, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 9437184 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = (xindex // 4718592) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x4 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x5 = xindex % 4718592 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_ptr0 + (2304 + x0), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr1 + (2304 + x0 + (6912*x2)), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tl.load(in_ptr2 + (2304 + x0), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tl.load(in_out_ptr0 + (x4), None).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tl.load(in_ptr3 + (x0), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tl.load(in_ptr4 + (x4), None).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = tl.load(in_ptr5 + (x0), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = tl.load(in_ptr6 + (x5), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tmp1 + tmp2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp0 + tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tmp5 + tmp6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp7 * tmp8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tmp4 * tmp9 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tmp11 + tmp12 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = tmp13 + tmp14 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = tmp10 + tmp15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(in_out_ptr0 + (x4), tmp16, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/gx/cgxt5of5zbfg7n4ls3hv3tgu5eykh3o4bgxsm3qo5ofijskh47uh.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_2], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # query_2 => constant_pad_nd_9 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10 = async_compile.triton('triton_poi_fused_constant_pad_nd_10', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16777216], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_10', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_10(in_ptr0, out_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 16777216 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 2048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 2048) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 1152, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1)), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp3, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/ym/cym22d2xvzw6lcvhvhh7s2xrku7mzzfhdhyvtnoorogbjt3rfjw6.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_3], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_3 => constant_pad_nd_4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_11 = async_compile.triton('triton_poi_fused_constant_pad_nd_11', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[2097152], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_11', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_11(in_ptr0, in_ptr1, out_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 1228800 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 2048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 2048) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 1152, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1) + (345600*((x1 % 300) // 300))), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tl.load(in_ptr1 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tmp3 + tmp4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tmp5.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = 0.5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tmp6 * tmp7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp6 * tmp6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tmp9 * tmp6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = 0.044715 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = tmp10 * tmp11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tmp6 + tmp12 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = 0.7978845608028654 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = tmp13 * tmp14 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = libdevice.tanh(tmp15) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp17 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18 = tmp16 + tmp17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19 = tmp8 * tmp18 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp20 = tmp19.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp21 = tl.full(tmp20.shape, 0.0, tmp20.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp22 = tl.where(tmp2, tmp20, tmp21) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp22, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/a6/ca6ajgsmiwk7mesiihdpcjbcgvi7mpvxztghix2rqcki4yfgvilo.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_2, key_6, value_2, value_6], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # key_2 => constant_pad_nd_10 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # key_6 => constant_pad_nd_21 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # value_2 => constant_pad_nd_11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # value_6 => constant_pad_nd_22 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12 = async_compile.triton('triton_poi_fused_constant_pad_nd_12', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[2097152], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_12', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_12(in_ptr0, in_ptr1, out_ptr0, out_ptr1, out_ptr2, out_ptr3, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 1228800 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 2048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 2048) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 1152, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1) + (345600*((x1 % 300) // 300))), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tl.load(in_ptr1 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tmp3 + tmp4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tl.full(tmp5.shape, 0.0, tmp5.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tl.where(tmp2, tmp5, tmp6) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp7, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr1 + (x2), tmp7, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr2 + (x2), tmp7, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr3 + (x2), tmp7, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/gr/cgrbxhvlchnp2nbuyeulry4myfybk7cilzyuhudqywuye235zggz.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_11 => _scaled_dot_product_efficient_attention V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13 = async_compile.triton('triton_poi_fused__scaled_dot_product_efficient_attention_13', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[1048576], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused__scaled_dot_product_efficient_attention_13', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused__scaled_dot_product_efficient_attention_13(in_out_ptr0, in_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 691200 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = xindex < xnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_out_ptr0 + (x2), xmask).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr0 + (x0), xmask, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 + tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(in_out_ptr0 + (x2), tmp2, xmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/vi/cvi62zxotjcncijppau52dxzd6rteio363jy2eac3cyt6qr4vvza.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11, hidden_states_30], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_11 => constant_pad_nd_12 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_30 => constant_pad_nd_23 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14 = async_compile.triton('triton_poi_fused_constant_pad_nd_14', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16384], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*i64', 1: '*bf16', 2: '*bf16', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_14', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 1, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_14(in_ptr0, out_ptr0, out_ptr1, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 9728 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = xindex < xnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 304 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = (xindex // 4864) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x3 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 300, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + (x0 + (300*x2)), tmp2 & xmask, eviction_policy='evict_last', other=0.0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp3.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tmp5 - tmp4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = -10000.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tmp6 * tmp7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tl.full(tmp8.shape, 0.0, tmp8.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tl.where(tmp2, tmp8, tmp9) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x3), tmp10, xmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr1 + (x3), tmp10, xmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/jn/cjntrr3amnv7csjvqrynnmbehsyli26bzls3getkhv64euuyd5jl.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_6, hidden_states_16, hidden_states_17, mul_6, norm_hidden_states_2, norm_hidden_states_3], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add_6 => add_24 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_16 => div_2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_17 => add_22 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # mul_6 => mul_17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_2 => add_23, convert_element_type_12, convert_element_type_13, mul_16, rsqrt_1, sub_2, var_mean_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_3 => add_25 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15 = async_compile.triton('triton_red_fused_add_div_mul_native_layer_norm_15', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.reduction( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[8192, 2048], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] reduction_hint=ReductionHint.INNER, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: 'i32', 8: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_div_mul_native_layer_norm_15', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 12, 'num_reduction': 2, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True} V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_red_fused_add_div_mul_native_layer_norm_15(in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, out_ptr2, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 8192 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rnumel = 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rbase = tl.arange(0, RBLOCK)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_mean = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_m2 = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_weight = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] for roffset in range(0, rnumel, RBLOCK): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rindex = roffset + rbase V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rmask = rindex < rnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] r1 = rindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_ptr0 + (r1 + (1152*x0)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr1 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tl.load(in_ptr2 + (r1 + (1152*x0)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 + tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp2 * tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tmp4 + tmp5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tmp6.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tl.broadcast_to(tmp7, [XBLOCK, RBLOCK]) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_mean_next, tmp9_m2_next, tmp9_weight_next = triton_helpers.welford_reduce( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8, tmp9_mean, tmp9_m2, tmp9_weight, roffset == 0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_mean = tl.where(rmask, tmp9_mean_next, tmp9_mean) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_m2 = tl.where(rmask, tmp9_m2_next, tmp9_m2) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_weight = tl.where(rmask, tmp9_weight_next, tmp9_weight) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_tmp, tmp10_tmp, tmp11_tmp = triton_helpers.welford( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9_mean, tmp9_m2, tmp9_weight, 1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp9_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tmp10_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tmp11_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x3 = (xindex // 4096) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] for roffset in range(0, rnumel, RBLOCK): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rindex = roffset + rbase V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rmask = rindex < rnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] r1 = rindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = tl.load(in_ptr0 + (r1 + (1152*x0)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tl.load(in_ptr1 + (r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp17 = tl.load(in_ptr2 + (r1 + (1152*x0)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp28 = tl.load(in_ptr3 + (4608 + r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp29 = tl.load(in_ptr4 + (4608 + r1 + (6912*x3)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp30 = tl.load(in_ptr5 + (4608 + r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp35 = tl.load(in_ptr3 + (3456 + r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp36 = tl.load(in_ptr4 + (3456 + r1 + (6912*x3)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp37 = tl.load(in_ptr5 + (3456 + r1), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = tmp12 + tmp13 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = tmp14 * tmp15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18 = tmp16 + tmp17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19 = tmp18.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp20 = tmp19 - tmp9 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp21 = 1152.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp22 = tmp10 / tmp21 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp23 = 1e-06 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp24 = tmp22 + tmp23 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp25 = libdevice.rsqrt(tmp24) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp26 = tmp20 * tmp25 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp27 = tmp26.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp31 = tmp29 + tmp30 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp32 = tmp28 + tmp31 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp33 = tmp32 + tmp15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp34 = tmp27 * tmp33 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp38 = tmp36 + tmp37 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp39 = tmp35 + tmp38 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp40 = tmp34 + tmp39 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr2 + (r1 + (1152*x0)), tmp40, rmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/kq/ckqn7d2kqemtlcbocxuabe6qk2zk2b5smq2wimkpq2klicoapdj3.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_21], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_21 => constant_pad_nd_15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16 = async_compile.triton('triton_poi_fused_constant_pad_nd_16', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[67108864], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_16', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_16(in_ptr0, in_ptr1, out_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 41943040 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 5120 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 5120) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 4608, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + (x0 + (4608*x1) + (18874368*((x1 % 4096) // 4096))), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tl.load(in_ptr1 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tmp3 + tmp4 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tmp5.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = 0.5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tmp6 * tmp7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp6 * tmp6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tmp9 * tmp6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = 0.044715 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = tmp10 * tmp11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tmp6 + tmp12 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = 0.7978845608028654 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = tmp13 * tmp14 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = libdevice.tanh(tmp15) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp17 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18 = tmp16 + tmp17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19 = tmp8 * tmp18 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp20 = tmp19.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp21 = tl.full(tmp20.shape, 0.0, tmp20.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp22 = tl.where(tmp2, tmp20, tmp21) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp22, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/ub/cuby5avq3u33nycjmmxevxnqqpmqsqqyi5tm5eh5ywizt5dixnbj.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_10, ff_output, hidden_states_16, hidden_states_17, hidden_states_22, mul_8, norm_hidden_states_4, norm_hidden_states_5], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # add_10 => add_33 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # ff_output => mul_24 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_16 => div_2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_17 => add_22 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_22 => add_30 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # mul_8 => mul_26 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_4 => add_32, convert_element_type_16, convert_element_type_17, mul_25, rsqrt_2, sub_3, var_mean_2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # norm_hidden_states_5 => add_34 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17 = async_compile.triton('triton_red_fused_add_div_mul_native_layer_norm_17', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.reduction( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[8192, 2048], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] reduction_hint=ReductionHint.INNER, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: '*bf16', 8: '*bf16', 9: '*bf16', 10: 'i32', 11: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_div_mul_native_layer_norm_17', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 15, 'num_reduction': 2, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True} V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_red_fused_add_div_mul_native_layer_norm_17(in_out_ptr0, in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, in_ptr6, in_ptr7, out_ptr2, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 8192 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rnumel = 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rbase = tl.arange(0, RBLOCK)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 4096) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x3 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_mean = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_m2 = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_weight = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] for roffset in range(0, rnumel, RBLOCK): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rindex = roffset + rbase V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rmask = rindex < rnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] r2 = rindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_ptr0 + (5760 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr1 + (5760 + r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tl.load(in_ptr2 + (5760 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tl.load(in_out_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tl.load(in_ptr3 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tl.load(in_ptr4 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tl.load(in_ptr5 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = tl.load(in_ptr6 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tmp1 + tmp2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp0 + tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tmp5 + tmp6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tmp4 * tmp7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tmp9 + tmp10 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tmp11 * tmp12 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = tmp13 + tmp14 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = tmp8 + tmp15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp17 = tmp16.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18 = tl.broadcast_to(tmp17, [XBLOCK, RBLOCK]) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_mean_next, tmp19_m2_next, tmp19_weight_next = triton_helpers.welford_reduce( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18, tmp19_mean, tmp19_m2, tmp19_weight, roffset == 0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_mean = tl.where(rmask, tmp19_mean_next, tmp19_mean) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_m2 = tl.where(rmask, tmp19_m2_next, tmp19_m2) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_weight = tl.where(rmask, tmp19_weight_next, tmp19_weight) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(in_out_ptr0 + (r2 + (1152*x3)), tmp16, rmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_tmp, tmp20_tmp, tmp21_tmp = triton_helpers.welford( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_mean, tmp19_m2, tmp19_weight, 1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19 = tmp19_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp20 = tmp20_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp21 = tmp21_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] for roffset in range(0, rnumel, RBLOCK): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rindex = roffset + rbase V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rmask = rindex < rnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] r2 = rindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp22 = tl.load(in_out_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp32 = tl.load(in_ptr7 + (1152 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp33 = tl.load(in_ptr1 + (1152 + r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp34 = tl.load(in_ptr2 + (1152 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp40 = tl.load(in_ptr7 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp41 = tl.load(in_ptr1 + (r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp42 = tl.load(in_ptr2 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp23 = tmp22.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp24 = tmp23 - tmp19 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp25 = 1152.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp26 = tmp20 / tmp25 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp27 = 1e-06 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp28 = tmp26 + tmp27 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp29 = libdevice.rsqrt(tmp28) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp30 = tmp24 * tmp29 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp31 = tmp30.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp35 = tmp33 + tmp34 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp36 = tmp32 + tmp35 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp37 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp38 = tmp36 + tmp37 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp39 = tmp31 * tmp38 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp43 = tmp41 + tmp42 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp44 = tmp40 + tmp43 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp45 = tmp39 + tmp44 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr2 + (r2 + (1152*x3)), tmp45, rmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/zl/czlcsoiyra2yl35d6vc2kjmszi72tek4i7g56lw2vawqn2tvlv63.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_1, hidden_states_28, hidden_states_29], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # attn_output_1 => mul_27 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_28 => div_3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_29 => add_39 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18 = async_compile.triton('triton_poi_fused_add_div_mul_18', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16777216], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_add_div_mul_18', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 6, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_add_div_mul_18(in_out_ptr0, in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 9437184 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = (xindex // 4718592) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x4 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_ptr0 + (2304 + x0), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr1 + (2304 + x0 + (6912*x2)), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tl.load(in_ptr2 + (2304 + x0), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tl.load(in_out_ptr0 + (x4), None).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tl.load(in_ptr3 + (x0), None, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tl.load(in_ptr4 + (x4), None).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tmp1 + tmp2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp0 + tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tmp5 + tmp6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp7 * tmp8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tmp4 * tmp9 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = tmp10 + tmp11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(in_out_ptr0 + (x4), tmp12, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/ks/ckskwpln7okqblhb2s5avhs6d6x67kskjaaoywrtq7pc7qyenaes.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [ff_output_27, hidden_states_529, hidden_states_530, hidden_states_535, hidden_states_536], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # ff_output_27 => mul_348 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_529 => div_56 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_530 => add_616 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_535 => add_624 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_536 => convert_element_type_178, var_mean_56 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_19 = async_compile.triton('triton_red_fused_add_div_mul_native_layer_norm_19', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.reduction( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[8192, 2048], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] reduction_hint=ReductionHint.INNER, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: '*bf16', 8: '*fp32', 9: '*fp32', 10: 'i32', 11: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_div_mul_native_layer_norm_19', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'num_load': 8, 'num_reduction': 2, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True} V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_red_fused_add_div_mul_native_layer_norm_19(in_out_ptr0, in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, in_ptr6, out_ptr0, out_ptr1, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 8192 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rnumel = 1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK, RBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rbase = tl.arange(0, RBLOCK)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 4096) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x3 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_mean = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_m2 = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_weight = tl.zeros([XBLOCK, RBLOCK], tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] for roffset in range(0, rnumel, RBLOCK): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rindex = roffset + rbase V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] rmask = rindex < rnumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] r2 = rindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_ptr0 + (5760 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr1 + (5760 + r2 + (6912*x1)), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tl.load(in_ptr2 + (5760 + r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tl.load(in_out_ptr0 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tl.load(in_ptr3 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tl.load(in_ptr4 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = tl.load(in_ptr5 + (r2), rmask, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = tl.load(in_ptr6 + (r2 + (1152*x3)), rmask, eviction_policy='evict_first', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tmp1 + tmp2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp0 + tmp3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tmp5 + tmp6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = tmp4 * tmp7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tmp9 + tmp10 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tmp11 * tmp12 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = tmp13 + tmp14 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = tmp8 + tmp15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp17 = tmp16.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18 = tl.broadcast_to(tmp17, [XBLOCK, RBLOCK]) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_mean_next, tmp19_m2_next, tmp19_weight_next = triton_helpers.welford_reduce( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18, tmp19_mean, tmp19_m2, tmp19_weight, roffset == 0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_mean = tl.where(rmask, tmp19_mean_next, tmp19_mean) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_m2 = tl.where(rmask, tmp19_m2_next, tmp19_m2) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_weight = tl.where(rmask, tmp19_weight_next, tmp19_weight) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(in_out_ptr0 + (r2 + (1152*x3)), tmp16, rmask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_tmp, tmp20_tmp, tmp21_tmp = triton_helpers.welford( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19_mean, tmp19_m2, tmp19_weight, 1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19 = tmp19_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp20 = tmp20_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp21 = tmp21_tmp[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x3), tmp19, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr1 + (x3), tmp20, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/yl/cylfrobtjz27y6zkonkaeawdjman53mf5gful2kwrjzkngczzxz3.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_538], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # hidden_states_538 => constant_pad_nd_313 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_20 = async_compile.triton('triton_poi_fused_constant_pad_nd_20', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16777216], V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*fp32', 2: '*fp32', 3: '*bf16', 4: '*bf16', 5: '*bf16', 6: '*bf16', 7: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4, 5, 6, 7), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_constant_pad_nd_20', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 7, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_constant_pad_nd_20(in_ptr0, in_ptr1, in_ptr2, in_ptr3, in_ptr4, in_ptr5, out_ptr0, xnumel, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 16777216 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x0 = xindex % 2048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x1 = (xindex // 2048) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = x0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.full([1], 1152, tl.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 < tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp3 = tl.load(in_ptr0 + (x0 + (1152*x1)), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp4 = tmp3.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp5 = tl.load(in_ptr1 + (x1), tmp2, eviction_policy='evict_last', other=0.0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp6 = tmp4 - tmp5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp7 = tl.load(in_ptr2 + (x1), tmp2, eviction_policy='evict_last', other=0.0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp8 = 1152.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp9 = tmp7 / tmp8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp10 = 1e-06 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp11 = tmp9 + tmp10 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp12 = libdevice.rsqrt(tmp11) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp13 = tmp6 * tmp12 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp14 = tmp13.to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp15 = tl.load(in_ptr3 + (1152 + x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp16 = tl.load(in_ptr4 + (x0 + (1152*(x1 // 4096))), tmp2, other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp17 = tl.load(in_ptr5 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp18 = tmp16 + tmp17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp19 = tmp15 + tmp18 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp20 = 1.0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp21 = tmp19 + tmp20 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp22 = tmp14 * tmp21 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp23 = tl.load(in_ptr3 + (x0), tmp2, eviction_policy='evict_last', other=0.0).to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp24 = tmp23 + tmp18 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp25 = tmp22 + tmp24 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp26 = tl.full(tmp25.shape, 0.0, tmp25.dtype) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp27 = tl.where(tmp2, tmp25, tmp26) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x2), tmp27, None) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # kernel path: /tmp/torchinductor_sayak/fk/cfksz3ijwl65jz33sewtbiwkc5sbwm7cr5juqhol74ejxyb4tijh.py V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [output], Original ATen: [aten.clone] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # output => clone_140 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_clone_21 = async_compile.triton('triton_poi_fused_clone_21', ''' V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] import triton.language as tl V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from triton.compiler.compiler import AttrsDescriptor V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime import triton_helpers, triton_heuristics V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.triton_helpers import libdevice, math as tl_math V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.runtime.hints import AutotuneHint, ReductionHint, TileHint, instance_descriptor, DeviceProperties V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton_heuristics.pointwise( V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] size_hints=[16, 16384], tile_hint=TileHint.DEFAULT, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] filename=__file__, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_meta={'signature': {0: '*bf16', 1: '*bf16', 2: '*bf16', 3: 'i32', 4: 'i32'}, 'device': DeviceProperties(type='cuda', index=0, cc=90, major=9, regs_per_multiprocessor=65536, max_threads_per_multi_processor=2048, multi_processor_count=132), 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3, 4), equal_to_1=())]}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] inductor_meta={'autotune_hints': set(), 'kernel_name': 'triton_poi_fused_clone_21', 'mutated_arg_names': [], 'no_x_dim': False, 'num_load': 2, 'num_reduction': 0, 'backend_hash': '002A1A9B1115CD8E0489B47343AA1BAA75B3F6181CDF90468122931EFBBE395F', 'are_deterministic_algorithms_enabled': False, 'assert_indirect_indexing': True, 'autotune_local_cache': True, 'autotune_pointwise': True, 'autotune_remote_cache': None, 'force_disable_caches': False, 'dynamic_scale_rblock': True, 'max_autotune': True, 'max_autotune_pointwise': False, 'min_split_scan_rblock': 256, 'spill_threshold': 16, 'store_cubin': False, 'coordinate_descent_tuning': True, 'coordinate_descent_search_radius': 1, 'coordinate_descent_check_all_directions': True}, V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] min_elem_per_thread=0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] @triton.jit V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def triton_poi_fused_clone_21(in_ptr0, in_ptr1, out_ptr0, ynumel, xnumel, YBLOCK : tl.constexpr, XBLOCK : tl.constexpr): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ynumel = 16 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xnumel = 16384 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] yoffset = tl.program_id(1) * YBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] yindex = yoffset + tl.arange(0, YBLOCK)[None, :] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ymask = yindex < ynumel V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xoffset = tl.program_id(0) * XBLOCK V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xindex = xoffset + tl.arange(0, XBLOCK)[:, None] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] xmask = tl.full([XBLOCK, YBLOCK], True, tl.int1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x2 = xindex % 2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x3 = (xindex // 2) % 64 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x4 = (xindex // 128) % 2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x5 = (xindex // 256) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] y0 = yindex % 8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] y1 = (yindex // 8) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] x7 = xindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] y6 = yindex V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp0 = tl.load(in_ptr0 + (y0 + (8*x2) + (16*x4) + (32*x3) + (2048*x5) + (131072*y1) + (131072*((x3 + (64*x5)) // 4096))), ymask, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp1 = tl.load(in_ptr1 + (y0 + (8*x2) + (16*x4)), ymask, eviction_policy='evict_last').to(tl.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tmp2 = tmp0 + tmp1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] tl.store(out_ptr0 + (x7 + (16384*y6)), tmp2, ymask) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] ''', device_str='cuda') V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] async_compile.wait(globals()) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del async_compile V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def call(args): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg0_1, arg1_1, arg2_1, arg3_1, arg4_1, arg5_1, arg6_1, arg7_1, arg8_1, arg9_1, arg10_1, arg11_1, arg12_1, arg13_1, arg14_1, arg15_1, arg16_1, arg17_1, arg18_1, arg19_1, arg20_1, arg21_1, arg22_1, arg23_1, arg24_1, arg25_1, arg26_1, arg27_1, arg28_1, arg29_1, arg30_1, arg31_1, arg32_1, arg33_1, arg34_1, arg35_1, arg36_1, arg37_1, arg38_1, arg39_1, arg40_1, arg41_1, arg42_1, arg43_1, arg44_1, arg45_1, arg46_1, arg47_1, arg48_1, arg49_1, arg50_1, arg51_1, arg52_1, arg53_1, arg54_1, arg55_1, arg56_1, arg57_1, arg58_1, arg59_1, arg60_1, arg61_1, arg62_1, arg63_1, arg64_1, arg65_1, arg66_1, arg67_1, arg68_1, arg69_1, arg70_1, arg71_1, arg72_1, arg73_1, arg74_1, arg75_1, arg76_1, arg77_1, arg78_1, arg79_1, arg80_1, arg81_1, arg82_1, arg83_1, arg84_1, arg85_1, arg86_1, arg87_1, arg88_1, arg89_1, arg90_1, arg91_1, arg92_1, arg93_1, arg94_1, arg95_1, arg96_1, arg97_1, arg98_1, arg99_1, arg100_1, arg101_1, arg102_1, arg103_1, arg104_1, arg105_1, arg106_1, arg107_1, arg108_1, arg109_1, arg110_1, arg111_1, arg112_1, arg113_1, arg114_1, arg115_1, arg116_1, arg117_1, arg118_1, arg119_1, arg120_1, arg121_1, arg122_1, arg123_1, arg124_1, arg125_1, arg126_1, arg127_1, arg128_1, arg129_1, arg130_1, arg131_1, arg132_1, arg133_1, arg134_1, arg135_1, arg136_1, arg137_1, arg138_1, arg139_1, arg140_1, arg141_1, arg142_1, arg143_1, arg144_1, arg145_1, arg146_1, arg147_1, arg148_1, arg149_1, arg150_1, arg151_1, arg152_1, arg153_1, arg154_1, arg155_1, arg156_1, arg157_1, arg158_1, arg159_1, arg160_1, arg161_1, arg162_1, arg163_1, arg164_1, arg165_1, arg166_1, arg167_1, arg168_1, arg169_1, arg170_1, arg171_1, arg172_1, arg173_1, arg174_1, arg175_1, arg176_1, arg177_1, arg178_1, arg179_1, arg180_1, arg181_1, arg182_1, arg183_1, arg184_1, arg185_1, arg186_1, arg187_1, arg188_1, arg189_1, arg190_1, arg191_1, arg192_1, arg193_1, arg194_1, arg195_1, arg196_1, arg197_1, arg198_1, arg199_1, arg200_1, arg201_1, arg202_1, arg203_1, arg204_1, arg205_1, arg206_1, arg207_1, arg208_1, arg209_1, arg210_1, arg211_1, arg212_1, arg213_1, arg214_1, arg215_1, arg216_1, arg217_1, arg218_1, arg219_1, arg220_1, arg221_1, arg222_1, arg223_1, arg224_1, arg225_1, arg226_1, arg227_1, arg228_1, arg229_1, arg230_1, arg231_1, arg232_1, arg233_1, arg234_1, arg235_1, arg236_1, arg237_1, arg238_1, arg239_1, arg240_1, arg241_1, arg242_1, arg243_1, arg244_1, arg245_1, arg246_1, arg247_1, arg248_1, arg249_1, arg250_1, arg251_1, arg252_1, arg253_1, arg254_1, arg255_1, arg256_1, arg257_1, arg258_1, arg259_1, arg260_1, arg261_1, arg262_1, arg263_1, arg264_1, arg265_1, arg266_1, arg267_1, arg268_1, arg269_1, arg270_1, arg271_1, arg272_1, arg273_1, arg274_1, arg275_1, arg276_1, arg277_1, arg278_1, arg279_1, arg280_1, arg281_1, arg282_1, arg283_1, arg284_1, arg285_1, arg286_1, arg287_1, arg288_1, arg289_1, arg290_1, arg291_1, arg292_1, arg293_1, arg294_1, arg295_1, arg296_1, arg297_1, arg298_1, arg299_1, arg300_1, arg301_1, arg302_1, arg303_1, arg304_1, arg305_1, arg306_1, arg307_1, arg308_1, arg309_1, arg310_1, arg311_1, arg312_1, arg313_1, arg314_1, arg315_1, arg316_1, arg317_1, arg318_1, arg319_1, arg320_1, arg321_1, arg322_1, arg323_1, arg324_1, arg325_1, arg326_1, arg327_1, arg328_1, arg329_1, arg330_1, arg331_1, arg332_1, arg333_1, arg334_1, arg335_1, arg336_1, arg337_1, arg338_1, arg339_1, arg340_1, arg341_1, arg342_1, arg343_1, arg344_1, arg345_1, arg346_1, arg347_1, arg348_1, arg349_1, arg350_1, arg351_1, arg352_1, arg353_1, arg354_1, arg355_1, arg356_1, arg357_1, arg358_1, arg359_1, arg360_1, arg361_1, arg362_1, arg363_1, arg364_1, arg365_1, arg366_1, arg367_1, arg368_1, arg369_1, arg370_1, arg371_1, arg372_1, arg373_1, arg374_1, arg375_1, arg376_1, arg377_1, arg378_1, arg379_1, arg380_1, arg381_1, arg382_1, arg383_1, arg384_1, arg385_1, arg386_1, arg387_1, arg388_1, arg389_1, arg390_1, arg391_1, arg392_1, arg393_1, arg394_1, arg395_1, arg396_1, arg397_1, arg398_1, arg399_1, arg400_1, arg401_1, arg402_1, arg403_1, arg404_1, arg405_1, arg406_1, arg407_1, arg408_1, arg409_1, arg410_1, arg411_1, arg412_1, arg413_1, arg414_1, arg415_1, arg416_1, arg417_1, arg418_1, arg419_1, arg420_1, arg421_1, arg422_1, arg423_1, arg424_1, arg425_1, arg426_1, arg427_1, arg428_1, arg429_1, arg430_1, arg431_1, arg432_1, arg433_1, arg434_1, arg435_1, arg436_1, arg437_1, arg438_1, arg439_1, arg440_1, arg441_1, arg442_1, arg443_1, arg444_1, arg445_1, arg446_1, arg447_1, arg448_1, arg449_1, arg450_1, arg451_1, arg452_1, arg453_1, arg454_1, arg455_1, arg456_1, arg457_1, arg458_1, arg459_1, arg460_1, arg461_1, arg462_1, arg463_1, arg464_1, arg465_1, arg466_1, arg467_1, arg468_1, arg469_1, arg470_1, arg471_1, arg472_1, arg473_1, arg474_1, arg475_1, arg476_1, arg477_1, arg478_1, arg479_1, arg480_1, arg481_1, arg482_1, arg483_1, arg484_1, arg485_1, arg486_1, arg487_1, arg488_1, arg489_1, arg490_1, arg491_1, arg492_1, arg493_1, arg494_1, arg495_1, arg496_1, arg497_1, arg498_1, arg499_1, arg500_1, arg501_1, arg502_1, arg503_1, arg504_1, arg505_1, arg506_1, arg507_1, arg508_1, arg509_1, arg510_1, arg511_1, arg512_1, arg513_1, arg514_1, arg515_1, arg516_1, arg517_1, arg518_1, arg519_1, arg520_1, arg521_1, arg522_1, arg523_1, arg524_1, arg525_1, arg526_1, arg527_1, arg528_1, arg529_1, arg530_1, arg531_1, arg532_1, arg533_1, arg534_1, arg535_1, arg536_1, arg537_1, arg538_1, arg539_1, arg540_1, arg541_1, arg542_1, arg543_1, arg544_1, arg545_1, arg546_1, arg547_1, arg548_1, arg549_1, arg550_1, arg551_1, arg552_1, arg553_1, arg554_1, arg555_1, arg556_1, arg557_1, arg558_1, arg559_1, arg560_1, arg561_1, arg562_1, arg563_1, arg564_1, arg565_1, arg566_1, arg567_1, arg568_1, arg569_1, arg570_1, arg571_1, arg572_1, arg573_1, arg574_1, arg575_1, arg576_1, arg577_1, arg578_1, arg579_1, arg580_1, arg581_1, arg582_1, arg583_1, arg584_1, arg585_1, arg586_1, arg587_1, arg588_1, arg589_1, arg590_1, arg591_1, arg592_1, arg593_1, arg594_1, arg595_1, arg596_1, arg597_1, arg598_1, arg599_1, arg600_1, arg601_1, arg602_1, arg603_1, arg604_1, arg605_1, arg606_1, arg607_1, arg608_1, arg609_1, arg610_1, arg611_1, arg612_1, arg613_1, arg614_1, arg615_1, arg616_1, arg617_1, arg618_1, arg619_1, arg620_1, arg621_1, arg622_1, arg623_1, arg624_1, arg625_1, arg626_1, arg627_1, arg628_1, arg629_1, arg630_1, arg631_1, arg632_1, arg633_1, arg634_1, arg635_1, arg636_1, arg637_1, arg638_1, arg639_1, arg640_1, arg641_1, arg642_1, arg643_1, arg644_1, arg645_1, arg646_1, arg647_1, arg648_1, arg649_1, arg650_1, arg651_1, arg652_1, arg653_1, arg654_1, arg655_1, arg656_1, arg657_1, arg658_1, arg659_1, arg660_1, arg661_1, arg662_1, arg663_1, arg664_1, arg665_1, arg666_1, arg667_1, arg668_1, arg669_1, arg670_1, arg671_1, arg672_1, arg673_1, arg674_1, arg675_1, arg676_1, arg677_1, arg678_1, arg679_1, arg680_1, arg681_1, arg682_1, arg683_1, arg684_1, arg685_1, arg686_1, arg687_1, arg688_1, arg689_1, arg690_1, arg691_1, arg692_1, arg693_1, arg694_1, arg695_1, arg696_1, arg697_1, arg698_1, arg699_1, arg700_1, arg701_1, arg702_1, arg703_1, arg704_1, arg705_1, arg706_1, arg707_1, arg708_1, arg709_1, arg710_1, arg711_1, arg712_1, arg713_1, arg714_1, arg715_1, arg716_1, arg717_1, arg718_1, arg719_1, arg720_1, arg721_1, arg722_1, arg723_1, arg724_1, arg725_1, arg726_1, arg727_1, arg728_1, arg729_1, arg730_1, arg731_1, arg732_1, arg733_1, arg734_1, arg735_1, arg736_1, arg737_1, arg738_1, arg739_1, arg740_1, arg741_1, arg742_1, arg743_1, arg744_1, arg745_1, arg746_1, arg747_1, arg748_1, arg749_1, arg750_1, arg751_1, arg752_1, arg753_1, arg754_1, arg755_1, arg756_1, arg757_1, arg758_1, arg759_1, arg760_1, arg761_1, arg762_1, arg763_1, arg764_1, arg765_1, arg766_1, arg767_1, arg768_1, arg769_1, arg770_1, arg771_1, arg772_1, arg773_1, arg774_1, arg775_1, arg776_1, arg777_1, arg778_1, arg779_1, arg780_1, arg781_1, arg782_1, arg783_1, arg784_1, arg785_1, arg786_1, arg787_1, arg788_1, arg789_1, arg790_1, arg791_1, arg792_1, arg793_1, arg794_1, arg795_1, arg796_1, arg797_1, arg798_1, arg799_1, arg800_1, arg801_1, arg802_1, arg803_1, arg804_1, arg805_1, arg806_1, arg807_1, arg808_1, arg809_1, arg810_1, arg811_1, arg812_1, arg813_1, arg814_1, arg815_1, arg816_1, arg817_1, arg818_1, arg819_1, arg820_1, arg821_1, arg822_1, arg823_1, arg824_1, arg825_1, arg826_1, arg827_1, arg828_1, arg829_1, arg830_1, arg831_1, arg832_1, arg833_1, arg834_1, arg835_1, arg836_1, arg837_1, arg838_1, arg839_1, arg840_1, arg841_1, arg842_1, arg843_1, arg844_1, arg845_1, arg846_1, arg847_1, arg848_1, arg849_1, arg850_1, arg851_1, arg852_1, arg853_1, arg854_1, arg855_1, arg856_1, arg857_1, arg858_1, arg859_1, arg860_1, arg861_1, arg862_1, arg863_1, arg864_1, arg865_1, arg866_1, arg867_1, arg868_1, arg869_1, arg870_1, arg871_1, arg872_1, arg873_1, arg874_1, arg875_1, arg876_1, arg877_1, arg878_1, arg879_1, arg880_1, arg881_1, arg882_1, arg883_1, arg884_1, arg885_1, arg886_1, arg887_1, arg888_1, arg889_1, arg890_1, arg891_1, arg892_1, arg893_1 = args V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] args.clear() V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg0_1, (2, 300), (300, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg1_1, (2, 4, 128, 128), (65536, 16384, 128, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg2_1, (1152, 4, 2, 2), (16, 1, 8, 4)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg3_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg4_1, (1, 4096, 1152), (4718592, 1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg5_1, (2, ), (0, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg6_1, (144, 8, 32, 4), (1024, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg7_1, (8, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg8_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg9_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg10_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg11_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg12_1, (864, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg13_1, (16, 6912, 2), (13824, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg14_1, (6912, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg15_1, (144, 32, 32, 4), (4096, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg16_1, (32, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg17_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg18_1, (2, 300, 4096), (1228800, 4096, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg19_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg20_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg21_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg22_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg23_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg24_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg25_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg26_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg27_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg28_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg29_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg30_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg31_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg32_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg33_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg34_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg35_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg36_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg37_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg38_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg39_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg40_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg41_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg42_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg43_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg44_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg45_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg46_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg47_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg48_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg49_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg50_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg51_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg52_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg53_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg54_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg55_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg56_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg57_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg58_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg59_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg60_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg61_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg62_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg63_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg64_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg65_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg66_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg67_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg68_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg69_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg70_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg71_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg72_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg73_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg74_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg75_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg76_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg77_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg78_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg79_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg80_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg81_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg82_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg83_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg84_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg85_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg86_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg87_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg88_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg89_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg90_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg91_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg92_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg93_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg94_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg95_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg96_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg97_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg98_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg99_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg100_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg101_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg102_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg103_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg104_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg105_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg106_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg107_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg108_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg109_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg110_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg111_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg112_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg113_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg114_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg115_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg116_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg117_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg118_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg119_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg120_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg121_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg122_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg123_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg124_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg125_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg126_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg127_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg128_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg129_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg130_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg131_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg132_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg133_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg134_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg135_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg136_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg137_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg138_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg139_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg140_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg141_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg142_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg143_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg144_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg145_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg146_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg147_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg148_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg149_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg150_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg151_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg152_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg153_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg154_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg155_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg156_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg157_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg158_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg159_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg160_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg161_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg162_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg163_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg164_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg165_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg166_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg167_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg168_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg169_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg170_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg171_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg172_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg173_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg174_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg175_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg176_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg177_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg178_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg179_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg180_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg181_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg182_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg183_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg184_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg185_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg186_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg187_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg188_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg189_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg190_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg191_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg192_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg193_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg194_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg195_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg196_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg197_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg198_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg199_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg200_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg201_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg202_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg203_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg204_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg205_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg206_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg207_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg208_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg209_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg210_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg211_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg212_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg213_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg214_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg215_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg216_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg217_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg218_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg219_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg220_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg221_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg222_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg223_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg224_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg225_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg226_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg227_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg228_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg229_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg230_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg231_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg232_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg233_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg234_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg235_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg236_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg237_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg238_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg239_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg240_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg241_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg242_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg243_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg244_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg245_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg246_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg247_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg248_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg249_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg250_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg251_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg252_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg253_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg254_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg255_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg256_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg257_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg258_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg259_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg260_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg261_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg262_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg263_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg264_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg265_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg266_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg267_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg268_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg269_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg270_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg271_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg272_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg273_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg274_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg275_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg276_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg277_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg278_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg279_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg280_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg281_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg282_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg283_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg284_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg285_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg286_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg287_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg288_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg289_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg290_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg291_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg292_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg293_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg294_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg295_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg296_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg297_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg298_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg299_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg300_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg301_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg302_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg303_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg304_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg305_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg306_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg307_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg308_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg309_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg310_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg311_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg312_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg313_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg314_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg315_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg316_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg317_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg318_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg319_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg320_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg321_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg322_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg323_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg324_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg325_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg326_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg327_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg328_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg329_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg330_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg331_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg332_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg333_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg334_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg335_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg336_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg337_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg338_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg339_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg340_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg341_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg342_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg343_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg344_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg345_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg346_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg347_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg348_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg349_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg350_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg351_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg352_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg353_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg354_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg355_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg356_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg357_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg358_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg359_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg360_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg361_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg362_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg363_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg364_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg365_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg366_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg367_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg368_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg369_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg370_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg371_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg372_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg373_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg374_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg375_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg376_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg377_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg378_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg379_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg380_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg381_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg382_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg383_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg384_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg385_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg386_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg387_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg388_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg389_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg390_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg391_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg392_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg393_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg394_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg395_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg396_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg397_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg398_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg399_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg400_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg401_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg402_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg403_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg404_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg405_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg406_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg407_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg408_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg409_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg410_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg411_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg412_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg413_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg414_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg415_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg416_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg417_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg418_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg419_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg420_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg421_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg422_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg423_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg424_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg425_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg426_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg427_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg428_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg429_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg430_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg431_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg432_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg433_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg434_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg435_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg436_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg437_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg438_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg439_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg440_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg441_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg442_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg443_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg444_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg445_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg446_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg447_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg448_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg449_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg450_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg451_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg452_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg453_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg454_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg455_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg456_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg457_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg458_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg459_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg460_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg461_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg462_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg463_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg464_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg465_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg466_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg467_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg468_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg469_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg470_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg471_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg472_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg473_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg474_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg475_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg476_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg477_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg478_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg479_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg480_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg481_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg482_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg483_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg484_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg485_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg486_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg487_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg488_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg489_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg490_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg491_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg492_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg493_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg494_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg495_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg496_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg497_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg498_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg499_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg500_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg501_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg502_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg503_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg504_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg505_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg506_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg507_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg508_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg509_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg510_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg511_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg512_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg513_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg514_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg515_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg516_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg517_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg518_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg519_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg520_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg521_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg522_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg523_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg524_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg525_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg526_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg527_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg528_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg529_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg530_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg531_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg532_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg533_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg534_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg535_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg536_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg537_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg538_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg539_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg540_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg541_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg542_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg543_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg544_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg545_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg546_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg547_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg548_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg549_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg550_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg551_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg552_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg553_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg554_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg555_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg556_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg557_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg558_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg559_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg560_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg561_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg562_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg563_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg564_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg565_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg566_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg567_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg568_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg569_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg570_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg571_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg572_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg573_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg574_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg575_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg576_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg577_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg578_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg579_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg580_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg581_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg582_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg583_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg584_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg585_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg586_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg587_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg588_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg589_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg590_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg591_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg592_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg593_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg594_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg595_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg596_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg597_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg598_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg599_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg600_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg601_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg602_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg603_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg604_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg605_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg606_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg607_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg608_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg609_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg610_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg611_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg612_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg613_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg614_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg615_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg616_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg617_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg618_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg619_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg620_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg621_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg622_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg623_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg624_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg625_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg626_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg627_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg628_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg629_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg630_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg631_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg632_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg633_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg634_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg635_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg636_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg637_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg638_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg639_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg640_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg641_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg642_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg643_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg644_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg645_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg646_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg647_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg648_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg649_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg650_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg651_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg652_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg653_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg654_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg655_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg656_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg657_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg658_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg659_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg660_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg661_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg662_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg663_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg664_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg665_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg666_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg667_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg668_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg669_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg670_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg671_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg672_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg673_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg674_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg675_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg676_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg677_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg678_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg679_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg680_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg681_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg682_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg683_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg684_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg685_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg686_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg687_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg688_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg689_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg690_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg691_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg692_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg693_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg694_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg695_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg696_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg697_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg698_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg699_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg700_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg701_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg702_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg703_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg704_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg705_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg706_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg707_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg708_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg709_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg710_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg711_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg712_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg713_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg714_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg715_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg716_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg717_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg718_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg719_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg720_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg721_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg722_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg723_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg724_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg725_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg726_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg727_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg728_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg729_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg730_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg731_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg732_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg733_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg734_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg735_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg736_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg737_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg738_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg739_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg740_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg741_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg742_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg743_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg744_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg745_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg746_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg747_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg748_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg749_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg750_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg751_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg752_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg753_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg754_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg755_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg756_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg757_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg758_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg759_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg760_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg761_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg762_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg763_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg764_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg765_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg766_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg767_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg768_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg769_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg770_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg771_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg772_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg773_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg774_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg775_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg776_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg777_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg778_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg779_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg780_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg781_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg782_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg783_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg784_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg785_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg786_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg787_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg788_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg789_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg790_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg791_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg792_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg793_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg794_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg795_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg796_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg797_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg798_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg799_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg800_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg801_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg802_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg803_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg804_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg805_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg806_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg807_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg808_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg809_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg810_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg811_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg812_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg813_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg814_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg815_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg816_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg817_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg818_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg819_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg820_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg821_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg822_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg823_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg824_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg825_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg826_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg827_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg828_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg829_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg830_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg831_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg832_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg833_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg834_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg835_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg836_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg837_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg838_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg839_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg840_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg841_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg842_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg843_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg844_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg845_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg846_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg847_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg848_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg849_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg850_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg851_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg852_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg853_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg854_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg855_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg856_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg857_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg858_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg859_1, (6, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg860_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg861_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg862_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg863_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg864_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg865_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg866_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg867_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg868_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg869_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg870_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg871_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg872_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg873_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg874_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg875_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg876_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg877_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg878_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg879_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg880_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg881_1, (144, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg882_1, (16, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg883_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg884_1, (576, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg885_1, (16, 4608, 2), (9216, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg886_1, (4608, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg887_1, (144, 40, 32, 4), (5120, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg888_1, (40, 1152, 2), (2304, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg889_1, (1152, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg890_1, (2, 1152), (1152, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg891_1, (4, 16, 32, 4), (2048, 128, 4, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg892_1, (16, 32, 2), (64, 2, 1)) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] assert_size_stride(arg893_1, (32, ), (1, )) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] with torch.cuda._DeviceGuard(0): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] torch.cuda.set_device(0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf0 = empty_strided_cuda((2, 256), (256, 1), torch.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [emb_3], Original ATen: [aten.cat] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] stream0 = get_raw_stream(0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_cat_0.run(arg5_1, buf0, 512, grid=grid(512), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg5_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1 = empty_strided_cuda((2, 1024), (1024, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [emb_4, sample, to_2], Original ATen: [aten._to_copy, aten.cat, aten.constant_pad_nd, aten.view] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__to_copy_cat_constant_pad_nd_view_1.run(buf0, buf1, 2048, grid=grid(2048), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf0 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [emb_4, sample, to_2], Original ATen: [aten._to_copy, aten._weight_int4pack_mm, aten.cat, aten.constant_pad_nd, aten.view] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf2 = torch.ops.aten._weight_int4pack_mm.default(buf1, arg6_1, 128, arg7_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg6_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg7_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf3 = buf2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf2 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf4 = empty_strided_cuda((2, 2048), (2048, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [sample, sample_1, sample_2], Original ATen: [aten.add, aten.constant_pad_nd, aten.silu, aten.view] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_constant_pad_nd_silu_view_2.run(buf3, arg8_1, buf4, 4096, grid=grid(4096), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg8_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf3 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [sample, sample_1, sample_2], Original ATen: [aten._weight_int4pack_mm, aten.add, aten.constant_pad_nd, aten.silu, aten.view] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf5 = torch.ops.aten._weight_int4pack_mm.default(buf4, arg9_1, 128, arg10_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg10_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg9_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf6 = buf5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf5 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf7 = buf4; del buf4 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [sample_2, silu_1, timestep], Original ATen: [aten.add, aten.constant_pad_nd, aten.silu, aten.view] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_constant_pad_nd_silu_view_2.run(buf6, arg11_1, buf7, 4096, grid=grid(4096), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [sample_2, silu_1, timestep], Original ATen: [aten._weight_int4pack_mm, aten.add, aten.constant_pad_nd, aten.silu, aten.view] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf8 = torch.ops.aten._weight_int4pack_mm.default(buf7, arg12_1, 128, arg13_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg12_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg13_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf7 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf9 = buf8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf8 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf10 = empty_strided_cuda((2, 4, 128, 128), (65536, 1, 512, 4), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [latent], Original ATen: [aten.convolution] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_convolution_3.run(arg1_1, buf10, 8, 16384, grid=grid(8, 16384), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg1_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf11 = empty_strided_cuda((2, 1152, 64, 64), (4718592, 1, 73728, 1152), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [latent], Original ATen: [aten.convolution] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_tem_fused_convolution_4.run(buf10, arg2_1, buf11, grid=torch._inductor.kernel.conv.conv2d_grid(2, 1152, 64, 64, meta0), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg2_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf10 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf15 = empty_strided_cuda((2, 4096, 1152), (4718592, 1152, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add, add_2, mul_4, norm_hidden_states, norm_hidden_states_1], Original ATen: [aten.add, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_mul_native_layer_norm_5.run(buf11, arg3_1, arg4_1, arg22_1, buf9, arg14_1, buf15, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf16 = empty_strided_cuda((8192, 2048), (2048, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf19 = empty_strided_cuda((8192, 2048), (2048, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf22 = empty_strided_cuda((8192, 2048), (2048, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key, query, value], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf15, buf16, buf19, buf22, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf15 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf17 = torch.ops.aten._weight_int4pack_mm.default(buf16, arg23_1, 128, arg24_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg23_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg24_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf18 = buf17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf17 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf20 = torch.ops.aten._weight_int4pack_mm.default(buf19, arg26_1, 128, arg27_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg26_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg27_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf21 = buf20 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf20 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf23 = torch.ops.aten._weight_int4pack_mm.default(buf22, arg29_1, 128, arg30_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg29_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg30_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf24 = buf23 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf23 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf25 = reinterpret_tensor(buf18, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf18 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf25, arg25_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg25_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf26 = reinterpret_tensor(buf21, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf21 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf26, arg28_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg28_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf27 = reinterpret_tensor(buf24, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf24 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf27, arg31_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg31_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_4], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf28 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf25, buf26, buf27, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf25 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf26 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf27 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf29 = buf28[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf28 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf33 = buf22; del buf22 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_7], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf29, buf33, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf29 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_7], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf34 = torch.ops.aten._weight_int4pack_mm.default(buf33, arg32_1, 128, arg33_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg32_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg33_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf35 = buf34 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf34 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf36 = reinterpret_tensor(buf35, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf35 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add, attn_output, hidden_states_10, hidden_states_9], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_9.run(buf36, arg22_1, buf9, arg14_1, arg34_1, buf11, arg3_1, arg4_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg34_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg3_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg4_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf11 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf37 = buf33; del buf33 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_2], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf36, buf37, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_2], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf38 = torch.ops.aten._weight_int4pack_mm.default(buf37, arg35_1, 128, arg36_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg35_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg36_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf39 = buf38 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf38 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_1], Original ATen: [aten._weight_int4pack_mm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf40 = torch.ops.aten._weight_int4pack_mm.default(reinterpret_tensor(arg18_1, (600, 4096), (4096, 1), 0), arg15_1, 128, arg16_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg15_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg16_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg18_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf41 = buf40 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf40 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf42 = empty_strided_cuda((600, 2048), (2048, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_3], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_11.run(buf41, arg17_1, buf42, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg17_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf41 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_3], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf43 = torch.ops.aten._weight_int4pack_mm.default(buf42, arg19_1, 128, arg20_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg19_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg20_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf44 = buf43 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf43 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf45 = buf42; del buf42 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf48 = empty_strided_cuda((600, 2048), (2048, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf102 = empty_strided_cuda((600, 2048), (2048, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf105 = empty_strided_cuda((600, 2048), (2048, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_2, key_6, value_2, value_6], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf45, buf48, buf102, buf105, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_2], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf46 = torch.ops.aten._weight_int4pack_mm.default(buf45, arg38_1, 128, arg39_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg38_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg39_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf47 = buf46 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf46 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_2], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf49 = torch.ops.aten._weight_int4pack_mm.default(buf48, arg41_1, 128, arg42_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg41_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg42_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf50 = buf49 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf49 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf51 = reinterpret_tensor(buf39, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf39 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf51, arg37_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg37_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf52 = reinterpret_tensor(buf47, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf47 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf52, arg40_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg40_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf53 = reinterpret_tensor(buf50, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf50 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf53, arg43_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg43_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf54 = empty_strided_cuda((2, 16, 1, 304), (4864, 304, 304, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf111 = empty_strided_cuda((2, 16, 1, 304), (4864, 304, 304, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11, hidden_states_30], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf54, buf111, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_11], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf55 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf51, buf52, buf53, reinterpret_tensor(buf54, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf51 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf52 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf53 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf56 = buf55[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf55 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf60 = buf37; del buf37 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_14], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf56, buf60, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_14], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf61 = torch.ops.aten._weight_int4pack_mm.default(buf60, arg44_1, 128, arg45_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg44_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg45_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf62 = buf61 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf61 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf66 = reinterpret_tensor(buf56, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf56 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_6, hidden_states_16, hidden_states_17, mul_6, norm_hidden_states_2, norm_hidden_states_3], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf62, arg46_1, buf36, arg22_1, buf9, arg14_1, buf66, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf67 = buf60; del buf60 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_18], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf66, buf67, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_18], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf68 = torch.ops.aten._weight_int4pack_mm.default(buf67, arg47_1, 128, arg48_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg47_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg48_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf69 = buf68 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf68 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf70 = empty_strided_cuda((8192, 5120), (5120, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_21], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf69, arg49_1, buf70, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg49_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf69 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_21], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf71 = torch.ops.aten._weight_int4pack_mm.default(buf70, arg50_1, 128, arg51_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg50_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg51_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf72 = buf71 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf71 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf73 = reinterpret_tensor(buf72, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf72 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf77 = buf66; del buf66 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_10, ff_output, hidden_states_16, hidden_states_17, hidden_states_22, mul_8, norm_hidden_states_4, norm_hidden_states_5], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf73, arg22_1, buf9, arg14_1, arg52_1, buf62, arg46_1, buf36, arg53_1, buf77, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg22_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg46_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg52_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf36 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf62 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf78 = buf67; del buf67 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf81 = buf19; del buf19 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf84 = buf16; del buf16 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_4, query_4, value_4], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf77, buf78, buf81, buf84, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf77 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_4], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf79 = torch.ops.aten._weight_int4pack_mm.default(buf78, arg54_1, 128, arg55_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg54_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg55_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf80 = buf79 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf79 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_4], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf82 = torch.ops.aten._weight_int4pack_mm.default(buf81, arg57_1, 128, arg58_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg57_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg58_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf83 = buf82 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf82 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_4], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf85 = torch.ops.aten._weight_int4pack_mm.default(buf84, arg60_1, 128, arg61_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg60_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg61_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf86 = buf85 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf85 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf87 = reinterpret_tensor(buf80, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf80 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_23], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf87, arg56_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg56_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf88 = reinterpret_tensor(buf83, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf83 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_23], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf88, arg59_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg59_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf89 = reinterpret_tensor(buf86, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf86 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_23], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf89, arg62_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg62_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_23], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf90 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf87, buf88, buf89, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf87 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf88 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf89 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf91 = buf90[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf90 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf95 = buf84; del buf84 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_26], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf91, buf95, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf91 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_26], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf96 = torch.ops.aten._weight_int4pack_mm.default(buf95, arg63_1, 128, arg64_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg63_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg64_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf97 = buf96 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf96 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf98 = reinterpret_tensor(buf97, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf97 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_1, hidden_states_28, hidden_states_29], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf98, arg53_1, buf9, arg14_1, arg65_1, buf73, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg65_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf73 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf99 = buf95; del buf95 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_6], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf98, buf99, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_6], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf100 = torch.ops.aten._weight_int4pack_mm.default(buf99, arg66_1, 128, arg67_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg66_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg67_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf101 = buf100 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf100 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_6], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf103 = torch.ops.aten._weight_int4pack_mm.default(buf102, arg69_1, 128, arg70_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg69_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg70_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf104 = buf103 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf103 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_6], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf106 = torch.ops.aten._weight_int4pack_mm.default(buf105, arg72_1, 128, arg73_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg72_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg73_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf107 = buf106 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf106 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf108 = reinterpret_tensor(buf101, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf101 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_30], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf108, arg68_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg68_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf109 = reinterpret_tensor(buf104, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf104 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_30], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf109, arg71_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg71_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf110 = reinterpret_tensor(buf107, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf107 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_30], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf110, arg74_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg74_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_30], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf112 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf108, buf109, buf110, reinterpret_tensor(buf111, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf108 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf109 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf110 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf113 = buf112[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf112 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf117 = buf99; del buf99 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_33], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf113, buf117, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_33], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf118 = torch.ops.aten._weight_int4pack_mm.default(buf117, arg75_1, 128, arg76_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg75_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg76_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf119 = buf118 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf118 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf123 = reinterpret_tensor(buf113, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf113 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_14, hidden_states_35, hidden_states_36, mul_10, norm_hidden_states_6, norm_hidden_states_7], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf119, arg77_1, buf98, arg53_1, buf9, arg14_1, buf123, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf124 = buf117; del buf117 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_37], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf123, buf124, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_37], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf125 = torch.ops.aten._weight_int4pack_mm.default(buf124, arg78_1, 128, arg79_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg78_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg79_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf126 = buf125 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf125 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf127 = buf70; del buf70 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_40], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf126, arg80_1, buf127, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg80_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf126 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_40], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf128 = torch.ops.aten._weight_int4pack_mm.default(buf127, arg81_1, 128, arg82_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg81_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg82_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf129 = buf128 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf128 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf130 = reinterpret_tensor(buf129, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf129 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf134 = buf123; del buf123 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_18, ff_output_1, hidden_states_35, hidden_states_36, hidden_states_41, mul_12, norm_hidden_states_8, norm_hidden_states_9], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf130, arg53_1, buf9, arg14_1, arg83_1, buf119, arg77_1, buf98, arg84_1, buf134, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg53_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg77_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg83_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf119 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf98 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf135 = buf124; del buf124 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf138 = buf81; del buf81 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf141 = buf78; del buf78 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_8, query_8, value_8], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf134, buf135, buf138, buf141, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf134 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_8], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf136 = torch.ops.aten._weight_int4pack_mm.default(buf135, arg85_1, 128, arg86_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg85_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg86_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf137 = buf136 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf136 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_8], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf139 = torch.ops.aten._weight_int4pack_mm.default(buf138, arg88_1, 128, arg89_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg88_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg89_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf140 = buf139 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf139 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_8], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf142 = torch.ops.aten._weight_int4pack_mm.default(buf141, arg91_1, 128, arg92_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg91_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg92_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf143 = buf142 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf142 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf144 = reinterpret_tensor(buf137, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf137 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_42], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf144, arg87_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg87_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf145 = reinterpret_tensor(buf140, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf140 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_42], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf145, arg90_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg90_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf146 = reinterpret_tensor(buf143, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf143 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_42], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf146, arg93_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg93_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_42], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf147 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf144, buf145, buf146, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf144 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf145 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf146 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf148 = buf147[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf147 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf152 = buf141; del buf141 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_45], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf148, buf152, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf148 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_45], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf153 = torch.ops.aten._weight_int4pack_mm.default(buf152, arg94_1, 128, arg95_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg94_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg95_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf154 = buf153 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf153 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf155 = reinterpret_tensor(buf154, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf154 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_2, hidden_states_47, hidden_states_48], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf155, arg84_1, buf9, arg14_1, arg96_1, buf130, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg96_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf130 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf156 = buf152; del buf152 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_10], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf155, buf156, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_10], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf157 = torch.ops.aten._weight_int4pack_mm.default(buf156, arg97_1, 128, arg98_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg97_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg98_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf158 = buf157 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf157 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf159 = buf105; del buf105 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf162 = buf102; del buf102 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf216 = buf48; del buf48 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf219 = buf45; del buf45 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_10, key_14, value_10, value_14], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf159, buf162, buf216, buf219, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_10], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf160 = torch.ops.aten._weight_int4pack_mm.default(buf159, arg100_1, 128, arg101_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg100_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg101_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf161 = buf160 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf160 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_10], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf163 = torch.ops.aten._weight_int4pack_mm.default(buf162, arg103_1, 128, arg104_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg103_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg104_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf164 = buf163 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf163 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf165 = reinterpret_tensor(buf158, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf158 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_49], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf165, arg99_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg99_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf166 = reinterpret_tensor(buf161, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf161 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_49], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf166, arg102_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg102_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf167 = reinterpret_tensor(buf164, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf164 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_49], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf167, arg105_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg105_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf168 = buf111; del buf111 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf225 = buf54; del buf54 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_49, hidden_states_68], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf168, buf225, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_49], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf169 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf165, buf166, buf167, reinterpret_tensor(buf168, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf165 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf166 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf167 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf170 = buf169[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf169 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf174 = buf156; del buf156 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_52], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf170, buf174, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_52], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf175 = torch.ops.aten._weight_int4pack_mm.default(buf174, arg106_1, 128, arg107_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg106_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg107_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf176 = buf175 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf175 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf180 = reinterpret_tensor(buf170, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf170 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_22, hidden_states_54, hidden_states_55, mul_14, norm_hidden_states_10, norm_hidden_states_11], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf176, arg108_1, buf155, arg84_1, buf9, arg14_1, buf180, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf181 = buf174; del buf174 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_56], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf180, buf181, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_56], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf182 = torch.ops.aten._weight_int4pack_mm.default(buf181, arg109_1, 128, arg110_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg109_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg110_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf183 = buf182 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf182 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf184 = buf127; del buf127 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_59], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf183, arg111_1, buf184, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg111_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf183 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_59], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf185 = torch.ops.aten._weight_int4pack_mm.default(buf184, arg112_1, 128, arg113_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg112_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg113_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf186 = buf185 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf185 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf187 = reinterpret_tensor(buf186, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf186 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf191 = buf180; del buf180 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_26, ff_output_2, hidden_states_54, hidden_states_55, hidden_states_60, mul_16, norm_hidden_states_12, norm_hidden_states_13], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf187, arg84_1, buf9, arg14_1, arg114_1, buf176, arg108_1, buf155, arg115_1, buf191, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg108_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg114_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg84_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf155 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf176 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf192 = buf181; del buf181 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf195 = buf138; del buf138 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf198 = buf135; del buf135 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_12, query_12, value_12], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf191, buf192, buf195, buf198, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf191 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_12], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf193 = torch.ops.aten._weight_int4pack_mm.default(buf192, arg116_1, 128, arg117_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg116_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg117_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf194 = buf193 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf193 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_12], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf196 = torch.ops.aten._weight_int4pack_mm.default(buf195, arg119_1, 128, arg120_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg119_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg120_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf197 = buf196 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf196 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_12], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf199 = torch.ops.aten._weight_int4pack_mm.default(buf198, arg122_1, 128, arg123_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg122_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg123_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf200 = buf199 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf199 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf201 = reinterpret_tensor(buf194, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf194 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_61], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf201, arg118_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg118_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf202 = reinterpret_tensor(buf197, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf197 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_61], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf202, arg121_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg121_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf203 = reinterpret_tensor(buf200, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf200 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_61], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf203, arg124_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg124_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_61], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf204 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf201, buf202, buf203, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf201 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf202 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf203 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf205 = buf204[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf204 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf209 = buf198; del buf198 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_64], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf205, buf209, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf205 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_64], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf210 = torch.ops.aten._weight_int4pack_mm.default(buf209, arg125_1, 128, arg126_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg125_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg126_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf211 = buf210 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf210 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf212 = reinterpret_tensor(buf211, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf211 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_3, hidden_states_66, hidden_states_67], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf212, arg115_1, buf9, arg14_1, arg127_1, buf187, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg127_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf187 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf213 = buf209; del buf209 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_14], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf212, buf213, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_14], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf214 = torch.ops.aten._weight_int4pack_mm.default(buf213, arg128_1, 128, arg129_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg128_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg129_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf215 = buf214 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf214 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_14], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf217 = torch.ops.aten._weight_int4pack_mm.default(buf216, arg131_1, 128, arg132_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg131_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg132_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf218 = buf217 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf217 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_14], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf220 = torch.ops.aten._weight_int4pack_mm.default(buf219, arg134_1, 128, arg135_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg134_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg135_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf221 = buf220 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf220 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf222 = reinterpret_tensor(buf215, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf215 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_68], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf222, arg130_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg130_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf223 = reinterpret_tensor(buf218, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf218 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_68], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf223, arg133_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg133_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf224 = reinterpret_tensor(buf221, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf221 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_68], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf224, arg136_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg136_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_68], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf226 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf222, buf223, buf224, reinterpret_tensor(buf225, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf222 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf223 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf224 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf227 = buf226[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf226 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf231 = buf213; del buf213 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_71], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf227, buf231, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_71], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf232 = torch.ops.aten._weight_int4pack_mm.default(buf231, arg137_1, 128, arg138_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg137_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg138_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf233 = buf232 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf232 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf237 = reinterpret_tensor(buf227, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf227 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_30, hidden_states_73, hidden_states_74, mul_18, norm_hidden_states_14, norm_hidden_states_15], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf233, arg139_1, buf212, arg115_1, buf9, arg14_1, buf237, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf238 = buf231; del buf231 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_75], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf237, buf238, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_75], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf239 = torch.ops.aten._weight_int4pack_mm.default(buf238, arg140_1, 128, arg141_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg140_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg141_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf240 = buf239 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf239 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf241 = buf184; del buf184 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_78], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf240, arg142_1, buf241, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg142_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf240 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_78], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf242 = torch.ops.aten._weight_int4pack_mm.default(buf241, arg143_1, 128, arg144_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg143_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg144_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf243 = buf242 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf242 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf244 = reinterpret_tensor(buf243, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf243 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf248 = buf237; del buf237 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_34, ff_output_3, hidden_states_73, hidden_states_74, hidden_states_79, mul_20, norm_hidden_states_16, norm_hidden_states_17], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf244, arg115_1, buf9, arg14_1, arg145_1, buf233, arg139_1, buf212, arg146_1, buf248, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg115_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg139_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg145_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf212 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf233 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf249 = buf238; del buf238 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf252 = buf195; del buf195 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf255 = buf192; del buf192 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_16, query_16, value_16], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf248, buf249, buf252, buf255, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf248 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_16], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf250 = torch.ops.aten._weight_int4pack_mm.default(buf249, arg147_1, 128, arg148_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg147_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg148_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf251 = buf250 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf250 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_16], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf253 = torch.ops.aten._weight_int4pack_mm.default(buf252, arg150_1, 128, arg151_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg150_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg151_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf254 = buf253 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf253 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_16], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf256 = torch.ops.aten._weight_int4pack_mm.default(buf255, arg153_1, 128, arg154_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg153_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg154_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf257 = buf256 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf256 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf258 = reinterpret_tensor(buf251, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf251 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_80], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf258, arg149_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg149_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf259 = reinterpret_tensor(buf254, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf254 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_80], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf259, arg152_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg152_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf260 = reinterpret_tensor(buf257, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf257 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_80], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf260, arg155_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg155_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_80], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf261 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf258, buf259, buf260, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf258 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf259 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf260 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf262 = buf261[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf261 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf266 = buf255; del buf255 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_83], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf262, buf266, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf262 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_83], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf267 = torch.ops.aten._weight_int4pack_mm.default(buf266, arg156_1, 128, arg157_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg156_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg157_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf268 = buf267 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf267 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf269 = reinterpret_tensor(buf268, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf268 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_4, hidden_states_85, hidden_states_86], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf269, arg146_1, buf9, arg14_1, arg158_1, buf244, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg158_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf244 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf270 = buf266; del buf266 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_18], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf269, buf270, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_18], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf271 = torch.ops.aten._weight_int4pack_mm.default(buf270, arg159_1, 128, arg160_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg159_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg160_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf272 = buf271 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf271 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf273 = buf219; del buf219 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf276 = buf216; del buf216 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf330 = buf162; del buf162 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf333 = buf159; del buf159 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_18, key_22, value_18, value_22], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf273, buf276, buf330, buf333, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_18], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf274 = torch.ops.aten._weight_int4pack_mm.default(buf273, arg162_1, 128, arg163_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg162_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg163_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf275 = buf274 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf274 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_18], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf277 = torch.ops.aten._weight_int4pack_mm.default(buf276, arg165_1, 128, arg166_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg165_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg166_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf278 = buf277 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf277 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf279 = reinterpret_tensor(buf272, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf272 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_87], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf279, arg161_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg161_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf280 = reinterpret_tensor(buf275, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf275 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_87], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf280, arg164_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg164_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf281 = reinterpret_tensor(buf278, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf278 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_87], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf281, arg167_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg167_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf282 = buf225; del buf225 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf339 = buf168; del buf168 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_106, hidden_states_87], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf282, buf339, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_87], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf283 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf279, buf280, buf281, reinterpret_tensor(buf282, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf279 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf280 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf281 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf284 = buf283[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf283 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf288 = buf270; del buf270 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_90], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf284, buf288, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_90], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf289 = torch.ops.aten._weight_int4pack_mm.default(buf288, arg168_1, 128, arg169_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg168_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg169_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf290 = buf289 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf289 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf294 = reinterpret_tensor(buf284, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf284 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_38, hidden_states_92, hidden_states_93, mul_22, norm_hidden_states_18, norm_hidden_states_19], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf290, arg170_1, buf269, arg146_1, buf9, arg14_1, buf294, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf295 = buf288; del buf288 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_94], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf294, buf295, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_94], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf296 = torch.ops.aten._weight_int4pack_mm.default(buf295, arg171_1, 128, arg172_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg171_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg172_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf297 = buf296 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf296 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf298 = buf241; del buf241 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_97], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf297, arg173_1, buf298, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg173_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf297 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_97], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf299 = torch.ops.aten._weight_int4pack_mm.default(buf298, arg174_1, 128, arg175_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg174_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg175_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf300 = buf299 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf299 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf301 = reinterpret_tensor(buf300, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf300 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf305 = buf294; del buf294 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_42, ff_output_4, hidden_states_92, hidden_states_93, hidden_states_98, mul_24, norm_hidden_states_20, norm_hidden_states_21], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf301, arg146_1, buf9, arg14_1, arg176_1, buf290, arg170_1, buf269, arg177_1, buf305, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg146_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg170_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg176_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf269 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf290 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf306 = buf295; del buf295 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf309 = buf252; del buf252 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf312 = buf249; del buf249 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_20, query_20, value_20], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf305, buf306, buf309, buf312, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf305 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_20], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf307 = torch.ops.aten._weight_int4pack_mm.default(buf306, arg178_1, 128, arg179_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg178_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg179_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf308 = buf307 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf307 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_20], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf310 = torch.ops.aten._weight_int4pack_mm.default(buf309, arg181_1, 128, arg182_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg181_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg182_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf311 = buf310 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf310 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_20], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf313 = torch.ops.aten._weight_int4pack_mm.default(buf312, arg184_1, 128, arg185_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg184_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg185_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf314 = buf313 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf313 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf315 = reinterpret_tensor(buf308, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf308 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_99], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf315, arg180_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg180_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf316 = reinterpret_tensor(buf311, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf311 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_99], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf316, arg183_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg183_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf317 = reinterpret_tensor(buf314, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf314 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_99], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf317, arg186_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg186_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_99], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf318 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf315, buf316, buf317, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf315 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf316 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf317 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf319 = buf318[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf318 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf323 = buf312; del buf312 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_102], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf319, buf323, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf319 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_102], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf324 = torch.ops.aten._weight_int4pack_mm.default(buf323, arg187_1, 128, arg188_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg187_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg188_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf325 = buf324 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf324 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf326 = reinterpret_tensor(buf325, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf325 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_5, hidden_states_104, hidden_states_105], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf326, arg177_1, buf9, arg14_1, arg189_1, buf301, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg189_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf301 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf327 = buf323; del buf323 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_22], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf326, buf327, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_22], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf328 = torch.ops.aten._weight_int4pack_mm.default(buf327, arg190_1, 128, arg191_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg190_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg191_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf329 = buf328 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf328 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_22], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf331 = torch.ops.aten._weight_int4pack_mm.default(buf330, arg193_1, 128, arg194_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg193_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg194_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf332 = buf331 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf331 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_22], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf334 = torch.ops.aten._weight_int4pack_mm.default(buf333, arg196_1, 128, arg197_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg196_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg197_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf335 = buf334 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf334 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf336 = reinterpret_tensor(buf329, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf329 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_106], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf336, arg192_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg192_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf337 = reinterpret_tensor(buf332, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf332 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_106], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf337, arg195_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg195_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf338 = reinterpret_tensor(buf335, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf335 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_106], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf338, arg198_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg198_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_106], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf340 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf336, buf337, buf338, reinterpret_tensor(buf339, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf336 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf337 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf338 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf341 = buf340[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf340 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf345 = buf327; del buf327 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_109], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf341, buf345, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_109], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf346 = torch.ops.aten._weight_int4pack_mm.default(buf345, arg199_1, 128, arg200_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg199_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg200_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf347 = buf346 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf346 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf351 = reinterpret_tensor(buf341, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf341 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_46, hidden_states_111, hidden_states_112, mul_26, norm_hidden_states_22, norm_hidden_states_23], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf347, arg201_1, buf326, arg177_1, buf9, arg14_1, buf351, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf352 = buf345; del buf345 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_113], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf351, buf352, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_113], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf353 = torch.ops.aten._weight_int4pack_mm.default(buf352, arg202_1, 128, arg203_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg202_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg203_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf354 = buf353 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf353 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf355 = buf298; del buf298 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_116], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf354, arg204_1, buf355, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg204_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf354 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_116], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf356 = torch.ops.aten._weight_int4pack_mm.default(buf355, arg205_1, 128, arg206_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg205_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg206_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf357 = buf356 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf356 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf358 = reinterpret_tensor(buf357, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf357 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf362 = buf351; del buf351 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_50, ff_output_5, hidden_states_111, hidden_states_112, hidden_states_117, mul_28, norm_hidden_states_24, norm_hidden_states_25], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf358, arg177_1, buf9, arg14_1, arg207_1, buf347, arg201_1, buf326, arg208_1, buf362, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg177_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg201_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg207_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf326 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf347 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf363 = buf352; del buf352 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf366 = buf309; del buf309 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf369 = buf306; del buf306 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_24, query_24, value_24], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf362, buf363, buf366, buf369, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf362 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_24], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf364 = torch.ops.aten._weight_int4pack_mm.default(buf363, arg209_1, 128, arg210_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg209_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg210_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf365 = buf364 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf364 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_24], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf367 = torch.ops.aten._weight_int4pack_mm.default(buf366, arg212_1, 128, arg213_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg212_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg213_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf368 = buf367 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf367 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_24], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf370 = torch.ops.aten._weight_int4pack_mm.default(buf369, arg215_1, 128, arg216_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg215_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg216_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf371 = buf370 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf370 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf372 = reinterpret_tensor(buf365, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf365 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_118], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf372, arg211_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg211_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf373 = reinterpret_tensor(buf368, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf368 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_118], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf373, arg214_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg214_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf374 = reinterpret_tensor(buf371, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf371 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_118], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf374, arg217_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg217_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_118], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf375 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf372, buf373, buf374, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf372 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf373 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf374 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf376 = buf375[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf375 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf380 = buf369; del buf369 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_121], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf376, buf380, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf376 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_121], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf381 = torch.ops.aten._weight_int4pack_mm.default(buf380, arg218_1, 128, arg219_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg218_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg219_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf382 = buf381 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf381 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf383 = reinterpret_tensor(buf382, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf382 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_6, hidden_states_123, hidden_states_124], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf383, arg208_1, buf9, arg14_1, arg220_1, buf358, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg220_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf358 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf384 = buf380; del buf380 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_26], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf383, buf384, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_26], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf385 = torch.ops.aten._weight_int4pack_mm.default(buf384, arg221_1, 128, arg222_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg221_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg222_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf386 = buf385 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf385 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf387 = buf333; del buf333 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf390 = buf330; del buf330 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf444 = buf276; del buf276 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf447 = buf273; del buf273 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_26, key_30, value_26, value_30], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf387, buf390, buf444, buf447, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_26], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf388 = torch.ops.aten._weight_int4pack_mm.default(buf387, arg224_1, 128, arg225_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg224_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg225_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf389 = buf388 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf388 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_26], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf391 = torch.ops.aten._weight_int4pack_mm.default(buf390, arg227_1, 128, arg228_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg227_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg228_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf392 = buf391 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf391 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf393 = reinterpret_tensor(buf386, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf386 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_125], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf393, arg223_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg223_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf394 = reinterpret_tensor(buf389, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf389 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_125], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf394, arg226_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg226_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf395 = reinterpret_tensor(buf392, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf392 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_125], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf395, arg229_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg229_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf396 = buf339; del buf339 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf453 = buf282; del buf282 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_125, hidden_states_144], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf396, buf453, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_125], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf397 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf393, buf394, buf395, reinterpret_tensor(buf396, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf393 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf394 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf395 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf398 = buf397[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf397 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf402 = buf384; del buf384 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_128], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf398, buf402, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_128], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf403 = torch.ops.aten._weight_int4pack_mm.default(buf402, arg230_1, 128, arg231_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg230_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg231_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf404 = buf403 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf403 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf408 = reinterpret_tensor(buf398, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf398 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_54, hidden_states_130, hidden_states_131, mul_30, norm_hidden_states_26, norm_hidden_states_27], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf404, arg232_1, buf383, arg208_1, buf9, arg14_1, buf408, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf409 = buf402; del buf402 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_132], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf408, buf409, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_132], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf410 = torch.ops.aten._weight_int4pack_mm.default(buf409, arg233_1, 128, arg234_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg233_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg234_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf411 = buf410 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf410 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf412 = buf355; del buf355 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_135], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf411, arg235_1, buf412, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg235_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf411 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_135], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf413 = torch.ops.aten._weight_int4pack_mm.default(buf412, arg236_1, 128, arg237_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg236_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg237_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf414 = buf413 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf413 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf415 = reinterpret_tensor(buf414, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf414 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf419 = buf408; del buf408 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_58, ff_output_6, hidden_states_130, hidden_states_131, hidden_states_136, mul_32, norm_hidden_states_28, norm_hidden_states_29], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf415, arg208_1, buf9, arg14_1, arg238_1, buf404, arg232_1, buf383, arg239_1, buf419, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg208_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg232_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg238_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf383 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf404 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf420 = buf409; del buf409 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf423 = buf366; del buf366 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf426 = buf363; del buf363 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_28, query_28, value_28], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf419, buf420, buf423, buf426, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf419 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_28], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf421 = torch.ops.aten._weight_int4pack_mm.default(buf420, arg240_1, 128, arg241_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg240_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg241_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf422 = buf421 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf421 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_28], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf424 = torch.ops.aten._weight_int4pack_mm.default(buf423, arg243_1, 128, arg244_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg243_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg244_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf425 = buf424 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf424 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_28], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf427 = torch.ops.aten._weight_int4pack_mm.default(buf426, arg246_1, 128, arg247_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg246_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg247_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf428 = buf427 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf427 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf429 = reinterpret_tensor(buf422, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf422 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_137], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf429, arg242_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg242_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf430 = reinterpret_tensor(buf425, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf425 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_137], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf430, arg245_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg245_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf431 = reinterpret_tensor(buf428, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf428 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_137], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf431, arg248_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg248_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_137], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf432 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf429, buf430, buf431, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf429 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf430 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf431 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf433 = buf432[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf432 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf437 = buf426; del buf426 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_140], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf433, buf437, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf433 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_140], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf438 = torch.ops.aten._weight_int4pack_mm.default(buf437, arg249_1, 128, arg250_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg249_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg250_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf439 = buf438 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf438 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf440 = reinterpret_tensor(buf439, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf439 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_7, hidden_states_142, hidden_states_143], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf440, arg239_1, buf9, arg14_1, arg251_1, buf415, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg251_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf415 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf441 = buf437; del buf437 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_30], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf440, buf441, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_30], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf442 = torch.ops.aten._weight_int4pack_mm.default(buf441, arg252_1, 128, arg253_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg252_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg253_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf443 = buf442 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf442 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_30], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf445 = torch.ops.aten._weight_int4pack_mm.default(buf444, arg255_1, 128, arg256_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg255_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg256_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf446 = buf445 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf445 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_30], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf448 = torch.ops.aten._weight_int4pack_mm.default(buf447, arg258_1, 128, arg259_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg258_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg259_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf449 = buf448 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf448 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf450 = reinterpret_tensor(buf443, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf443 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_144], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf450, arg254_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg254_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf451 = reinterpret_tensor(buf446, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf446 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_144], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf451, arg257_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg257_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf452 = reinterpret_tensor(buf449, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf449 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_144], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf452, arg260_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg260_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_144], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf454 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf450, buf451, buf452, reinterpret_tensor(buf453, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf450 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf451 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf452 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf455 = buf454[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf454 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf459 = buf441; del buf441 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_147], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf455, buf459, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_147], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf460 = torch.ops.aten._weight_int4pack_mm.default(buf459, arg261_1, 128, arg262_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg261_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg262_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf461 = buf460 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf460 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf465 = reinterpret_tensor(buf455, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf455 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_62, hidden_states_149, hidden_states_150, mul_34, norm_hidden_states_30, norm_hidden_states_31], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf461, arg263_1, buf440, arg239_1, buf9, arg14_1, buf465, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf466 = buf459; del buf459 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_151], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf465, buf466, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_151], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf467 = torch.ops.aten._weight_int4pack_mm.default(buf466, arg264_1, 128, arg265_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg264_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg265_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf468 = buf467 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf467 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf469 = buf412; del buf412 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_154], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf468, arg266_1, buf469, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg266_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf468 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_154], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf470 = torch.ops.aten._weight_int4pack_mm.default(buf469, arg267_1, 128, arg268_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg267_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg268_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf471 = buf470 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf470 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf472 = reinterpret_tensor(buf471, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf471 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf476 = buf465; del buf465 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_66, ff_output_7, hidden_states_149, hidden_states_150, hidden_states_155, mul_36, norm_hidden_states_32, norm_hidden_states_33], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf472, arg239_1, buf9, arg14_1, arg269_1, buf461, arg263_1, buf440, arg270_1, buf476, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg239_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg263_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg269_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf440 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf461 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf477 = buf466; del buf466 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf480 = buf423; del buf423 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf483 = buf420; del buf420 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_32, query_32, value_32], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf476, buf477, buf480, buf483, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf476 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_32], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf478 = torch.ops.aten._weight_int4pack_mm.default(buf477, arg271_1, 128, arg272_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg271_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg272_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf479 = buf478 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf478 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_32], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf481 = torch.ops.aten._weight_int4pack_mm.default(buf480, arg274_1, 128, arg275_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg274_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg275_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf482 = buf481 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf481 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_32], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf484 = torch.ops.aten._weight_int4pack_mm.default(buf483, arg277_1, 128, arg278_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg277_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg278_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf485 = buf484 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf484 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf486 = reinterpret_tensor(buf479, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf479 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_156], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf486, arg273_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg273_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf487 = reinterpret_tensor(buf482, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf482 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_156], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf487, arg276_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg276_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf488 = reinterpret_tensor(buf485, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf485 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_156], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf488, arg279_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg279_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_156], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf489 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf486, buf487, buf488, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf486 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf487 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf488 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf490 = buf489[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf489 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf494 = buf483; del buf483 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_159], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf490, buf494, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf490 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_159], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf495 = torch.ops.aten._weight_int4pack_mm.default(buf494, arg280_1, 128, arg281_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg280_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg281_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf496 = buf495 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf495 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf497 = reinterpret_tensor(buf496, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf496 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_8, hidden_states_161, hidden_states_162], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf497, arg270_1, buf9, arg14_1, arg282_1, buf472, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg282_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf472 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf498 = buf494; del buf494 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_34], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf497, buf498, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_34], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf499 = torch.ops.aten._weight_int4pack_mm.default(buf498, arg283_1, 128, arg284_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg283_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg284_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf500 = buf499 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf499 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf501 = buf447; del buf447 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf504 = buf444; del buf444 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf558 = buf390; del buf390 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf561 = buf387; del buf387 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_34, key_38, value_34, value_38], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf501, buf504, buf558, buf561, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_34], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf502 = torch.ops.aten._weight_int4pack_mm.default(buf501, arg286_1, 128, arg287_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg286_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg287_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf503 = buf502 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf502 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_34], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf505 = torch.ops.aten._weight_int4pack_mm.default(buf504, arg289_1, 128, arg290_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg289_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg290_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf506 = buf505 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf505 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf507 = reinterpret_tensor(buf500, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf500 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_163], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf507, arg285_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg285_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf508 = reinterpret_tensor(buf503, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf503 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_163], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf508, arg288_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg288_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf509 = reinterpret_tensor(buf506, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf506 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_163], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf509, arg291_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg291_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf510 = buf453; del buf453 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf567 = buf396; del buf396 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_163, hidden_states_182], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf510, buf567, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_163], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf511 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf507, buf508, buf509, reinterpret_tensor(buf510, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf507 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf508 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf509 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf512 = buf511[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf511 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf516 = buf498; del buf498 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_166], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf512, buf516, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_166], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf517 = torch.ops.aten._weight_int4pack_mm.default(buf516, arg292_1, 128, arg293_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg292_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg293_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf518 = buf517 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf517 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf522 = reinterpret_tensor(buf512, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf512 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_70, hidden_states_168, hidden_states_169, mul_38, norm_hidden_states_34, norm_hidden_states_35], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf518, arg294_1, buf497, arg270_1, buf9, arg14_1, buf522, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf523 = buf516; del buf516 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_170], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf522, buf523, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_170], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf524 = torch.ops.aten._weight_int4pack_mm.default(buf523, arg295_1, 128, arg296_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg295_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg296_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf525 = buf524 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf524 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf526 = buf469; del buf469 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_173], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf525, arg297_1, buf526, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg297_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf525 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_173], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf527 = torch.ops.aten._weight_int4pack_mm.default(buf526, arg298_1, 128, arg299_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg298_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg299_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf528 = buf527 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf527 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf529 = reinterpret_tensor(buf528, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf528 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf533 = buf522; del buf522 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_74, ff_output_8, hidden_states_168, hidden_states_169, hidden_states_174, mul_40, norm_hidden_states_36, norm_hidden_states_37], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf529, arg270_1, buf9, arg14_1, arg300_1, buf518, arg294_1, buf497, arg301_1, buf533, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg270_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg294_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg300_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf497 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf518 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf534 = buf523; del buf523 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf537 = buf480; del buf480 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf540 = buf477; del buf477 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_36, query_36, value_36], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf533, buf534, buf537, buf540, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf533 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_36], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf535 = torch.ops.aten._weight_int4pack_mm.default(buf534, arg302_1, 128, arg303_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg302_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg303_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf536 = buf535 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf535 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_36], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf538 = torch.ops.aten._weight_int4pack_mm.default(buf537, arg305_1, 128, arg306_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg305_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg306_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf539 = buf538 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf538 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_36], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf541 = torch.ops.aten._weight_int4pack_mm.default(buf540, arg308_1, 128, arg309_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg308_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg309_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf542 = buf541 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf541 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf543 = reinterpret_tensor(buf536, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf536 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_175], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf543, arg304_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg304_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf544 = reinterpret_tensor(buf539, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf539 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_175], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf544, arg307_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg307_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf545 = reinterpret_tensor(buf542, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf542 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_175], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf545, arg310_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg310_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_175], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf546 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf543, buf544, buf545, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf543 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf544 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf545 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf547 = buf546[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf546 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf551 = buf540; del buf540 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_178], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf547, buf551, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf547 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_178], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf552 = torch.ops.aten._weight_int4pack_mm.default(buf551, arg311_1, 128, arg312_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg311_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg312_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf553 = buf552 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf552 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf554 = reinterpret_tensor(buf553, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf553 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_9, hidden_states_180, hidden_states_181], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf554, arg301_1, buf9, arg14_1, arg313_1, buf529, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg313_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf529 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf555 = buf551; del buf551 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_38], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf554, buf555, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_38], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf556 = torch.ops.aten._weight_int4pack_mm.default(buf555, arg314_1, 128, arg315_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg314_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg315_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf557 = buf556 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf556 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_38], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf559 = torch.ops.aten._weight_int4pack_mm.default(buf558, arg317_1, 128, arg318_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg317_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg318_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf560 = buf559 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf559 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_38], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf562 = torch.ops.aten._weight_int4pack_mm.default(buf561, arg320_1, 128, arg321_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg320_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg321_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf563 = buf562 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf562 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf564 = reinterpret_tensor(buf557, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf557 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_182], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf564, arg316_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg316_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf565 = reinterpret_tensor(buf560, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf560 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_182], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf565, arg319_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg319_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf566 = reinterpret_tensor(buf563, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf563 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_182], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf566, arg322_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg322_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_182], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf568 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf564, buf565, buf566, reinterpret_tensor(buf567, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf564 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf565 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf566 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf569 = buf568[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf568 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf573 = buf555; del buf555 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_185], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf569, buf573, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_185], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf574 = torch.ops.aten._weight_int4pack_mm.default(buf573, arg323_1, 128, arg324_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg323_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg324_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf575 = buf574 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf574 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf579 = reinterpret_tensor(buf569, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf569 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_78, hidden_states_187, hidden_states_188, mul_42, norm_hidden_states_38, norm_hidden_states_39], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf575, arg325_1, buf554, arg301_1, buf9, arg14_1, buf579, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf580 = buf573; del buf573 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_189], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf579, buf580, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_189], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf581 = torch.ops.aten._weight_int4pack_mm.default(buf580, arg326_1, 128, arg327_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg326_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg327_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf582 = buf581 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf581 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf583 = buf526; del buf526 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_192], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf582, arg328_1, buf583, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg328_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf582 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_192], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf584 = torch.ops.aten._weight_int4pack_mm.default(buf583, arg329_1, 128, arg330_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg329_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg330_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf585 = buf584 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf584 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf586 = reinterpret_tensor(buf585, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf585 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf590 = buf579; del buf579 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_82, ff_output_9, hidden_states_187, hidden_states_188, hidden_states_193, mul_44, norm_hidden_states_40, norm_hidden_states_41], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf586, arg301_1, buf9, arg14_1, arg331_1, buf575, arg325_1, buf554, arg332_1, buf590, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg301_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg325_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg331_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf554 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf575 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf591 = buf580; del buf580 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf594 = buf537; del buf537 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf597 = buf534; del buf534 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_40, query_40, value_40], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf590, buf591, buf594, buf597, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf590 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_40], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf592 = torch.ops.aten._weight_int4pack_mm.default(buf591, arg333_1, 128, arg334_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg333_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg334_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf593 = buf592 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf592 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_40], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf595 = torch.ops.aten._weight_int4pack_mm.default(buf594, arg336_1, 128, arg337_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg336_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg337_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf596 = buf595 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf595 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_40], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf598 = torch.ops.aten._weight_int4pack_mm.default(buf597, arg339_1, 128, arg340_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg339_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg340_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf599 = buf598 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf598 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf600 = reinterpret_tensor(buf593, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf593 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_194], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf600, arg335_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg335_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf601 = reinterpret_tensor(buf596, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf596 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_194], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf601, arg338_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg338_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf602 = reinterpret_tensor(buf599, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf599 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_194], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf602, arg341_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg341_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_194], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf603 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf600, buf601, buf602, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf600 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf601 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf602 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf604 = buf603[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf603 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf608 = buf597; del buf597 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_197], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf604, buf608, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf604 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_197], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf609 = torch.ops.aten._weight_int4pack_mm.default(buf608, arg342_1, 128, arg343_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg342_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg343_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf610 = buf609 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf609 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf611 = reinterpret_tensor(buf610, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf610 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_10, hidden_states_199, hidden_states_200], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf611, arg332_1, buf9, arg14_1, arg344_1, buf586, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg344_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf586 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf612 = buf608; del buf608 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_42], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf611, buf612, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_42], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf613 = torch.ops.aten._weight_int4pack_mm.default(buf612, arg345_1, 128, arg346_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg345_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg346_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf614 = buf613 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf613 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf615 = buf561; del buf561 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf618 = buf558; del buf558 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf672 = buf504; del buf504 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf675 = buf501; del buf501 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_42, key_46, value_42, value_46], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf615, buf618, buf672, buf675, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_42], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf616 = torch.ops.aten._weight_int4pack_mm.default(buf615, arg348_1, 128, arg349_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg348_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg349_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf617 = buf616 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf616 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_42], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf619 = torch.ops.aten._weight_int4pack_mm.default(buf618, arg351_1, 128, arg352_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg351_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg352_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf620 = buf619 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf619 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf621 = reinterpret_tensor(buf614, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf614 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_201], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf621, arg347_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg347_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf622 = reinterpret_tensor(buf617, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf617 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_201], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf622, arg350_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg350_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf623 = reinterpret_tensor(buf620, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf620 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_201], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf623, arg353_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg353_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf624 = buf567; del buf567 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf681 = buf510; del buf510 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_201, hidden_states_220], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf624, buf681, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_201], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf625 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf621, buf622, buf623, reinterpret_tensor(buf624, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf621 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf622 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf623 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf626 = buf625[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf625 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf630 = buf612; del buf612 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_204], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf626, buf630, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_204], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf631 = torch.ops.aten._weight_int4pack_mm.default(buf630, arg354_1, 128, arg355_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg354_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg355_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf632 = buf631 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf631 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf636 = reinterpret_tensor(buf626, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf626 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_86, hidden_states_206, hidden_states_207, mul_46, norm_hidden_states_42, norm_hidden_states_43], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf632, arg356_1, buf611, arg332_1, buf9, arg14_1, buf636, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf637 = buf630; del buf630 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_208], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf636, buf637, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_208], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf638 = torch.ops.aten._weight_int4pack_mm.default(buf637, arg357_1, 128, arg358_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg357_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg358_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf639 = buf638 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf638 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf640 = buf583; del buf583 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_211], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf639, arg359_1, buf640, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg359_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf639 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_211], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf641 = torch.ops.aten._weight_int4pack_mm.default(buf640, arg360_1, 128, arg361_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg360_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg361_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf642 = buf641 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf641 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf643 = reinterpret_tensor(buf642, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf642 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf647 = buf636; del buf636 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_90, ff_output_10, hidden_states_206, hidden_states_207, hidden_states_212, mul_48, norm_hidden_states_44, norm_hidden_states_45], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf643, arg332_1, buf9, arg14_1, arg362_1, buf632, arg356_1, buf611, arg363_1, buf647, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg332_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg356_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg362_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf611 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf632 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf648 = buf637; del buf637 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf651 = buf594; del buf594 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf654 = buf591; del buf591 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_44, query_44, value_44], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf647, buf648, buf651, buf654, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf647 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_44], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf649 = torch.ops.aten._weight_int4pack_mm.default(buf648, arg364_1, 128, arg365_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg364_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg365_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf650 = buf649 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf649 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_44], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf652 = torch.ops.aten._weight_int4pack_mm.default(buf651, arg367_1, 128, arg368_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg367_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg368_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf653 = buf652 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf652 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_44], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf655 = torch.ops.aten._weight_int4pack_mm.default(buf654, arg370_1, 128, arg371_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg370_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg371_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf656 = buf655 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf655 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf657 = reinterpret_tensor(buf650, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf650 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_213], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf657, arg366_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg366_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf658 = reinterpret_tensor(buf653, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf653 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_213], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf658, arg369_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg369_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf659 = reinterpret_tensor(buf656, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf656 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_213], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf659, arg372_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg372_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_213], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf660 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf657, buf658, buf659, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf657 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf658 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf659 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf661 = buf660[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf660 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf665 = buf654; del buf654 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_216], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf661, buf665, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf661 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_216], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf666 = torch.ops.aten._weight_int4pack_mm.default(buf665, arg373_1, 128, arg374_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg373_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg374_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf667 = buf666 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf666 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf668 = reinterpret_tensor(buf667, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf667 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_11, hidden_states_218, hidden_states_219], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf668, arg363_1, buf9, arg14_1, arg375_1, buf643, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg375_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf643 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf669 = buf665; del buf665 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_46], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf668, buf669, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_46], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf670 = torch.ops.aten._weight_int4pack_mm.default(buf669, arg376_1, 128, arg377_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg376_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg377_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf671 = buf670 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf670 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_46], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf673 = torch.ops.aten._weight_int4pack_mm.default(buf672, arg379_1, 128, arg380_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg379_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg380_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf674 = buf673 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf673 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_46], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf676 = torch.ops.aten._weight_int4pack_mm.default(buf675, arg382_1, 128, arg383_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg382_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg383_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf677 = buf676 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf676 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf678 = reinterpret_tensor(buf671, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf671 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_220], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf678, arg378_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg378_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf679 = reinterpret_tensor(buf674, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf674 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_220], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf679, arg381_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg381_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf680 = reinterpret_tensor(buf677, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf677 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_220], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf680, arg384_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg384_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_220], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf682 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf678, buf679, buf680, reinterpret_tensor(buf681, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf678 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf679 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf680 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf683 = buf682[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf682 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf687 = buf669; del buf669 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_223], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf683, buf687, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_223], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf688 = torch.ops.aten._weight_int4pack_mm.default(buf687, arg385_1, 128, arg386_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg385_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg386_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf689 = buf688 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf688 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf693 = reinterpret_tensor(buf683, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf683 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_94, hidden_states_225, hidden_states_226, mul_50, norm_hidden_states_46, norm_hidden_states_47], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf689, arg387_1, buf668, arg363_1, buf9, arg14_1, buf693, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf694 = buf687; del buf687 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_227], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf693, buf694, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_227], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf695 = torch.ops.aten._weight_int4pack_mm.default(buf694, arg388_1, 128, arg389_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg388_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg389_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf696 = buf695 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf695 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf697 = buf640; del buf640 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_230], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf696, arg390_1, buf697, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg390_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf696 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_230], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf698 = torch.ops.aten._weight_int4pack_mm.default(buf697, arg391_1, 128, arg392_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg391_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg392_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf699 = buf698 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf698 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf700 = reinterpret_tensor(buf699, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf699 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf704 = buf693; del buf693 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_98, ff_output_11, hidden_states_225, hidden_states_226, hidden_states_231, mul_52, norm_hidden_states_48, norm_hidden_states_49], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf700, arg363_1, buf9, arg14_1, arg393_1, buf689, arg387_1, buf668, arg394_1, buf704, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg363_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg387_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg393_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf668 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf689 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf705 = buf694; del buf694 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf708 = buf651; del buf651 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf711 = buf648; del buf648 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_48, query_48, value_48], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf704, buf705, buf708, buf711, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf704 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_48], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf706 = torch.ops.aten._weight_int4pack_mm.default(buf705, arg395_1, 128, arg396_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg395_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg396_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf707 = buf706 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf706 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_48], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf709 = torch.ops.aten._weight_int4pack_mm.default(buf708, arg398_1, 128, arg399_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg398_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg399_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf710 = buf709 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf709 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_48], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf712 = torch.ops.aten._weight_int4pack_mm.default(buf711, arg401_1, 128, arg402_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg401_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg402_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf713 = buf712 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf712 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf714 = reinterpret_tensor(buf707, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf707 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_232], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf714, arg397_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg397_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf715 = reinterpret_tensor(buf710, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf710 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_232], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf715, arg400_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg400_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf716 = reinterpret_tensor(buf713, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf713 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_232], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf716, arg403_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg403_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_232], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf717 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf714, buf715, buf716, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf714 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf715 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf716 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf718 = buf717[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf717 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf722 = buf711; del buf711 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_235], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf718, buf722, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf718 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_235], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf723 = torch.ops.aten._weight_int4pack_mm.default(buf722, arg404_1, 128, arg405_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg404_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg405_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf724 = buf723 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf723 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf725 = reinterpret_tensor(buf724, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf724 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_12, hidden_states_237, hidden_states_238], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf725, arg394_1, buf9, arg14_1, arg406_1, buf700, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg406_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf700 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf726 = buf722; del buf722 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_50], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf725, buf726, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_50], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf727 = torch.ops.aten._weight_int4pack_mm.default(buf726, arg407_1, 128, arg408_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg407_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg408_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf728 = buf727 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf727 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf729 = buf675; del buf675 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf732 = buf672; del buf672 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf786 = buf618; del buf618 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf789 = buf615; del buf615 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_50, key_54, value_50, value_54], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf729, buf732, buf786, buf789, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_50], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf730 = torch.ops.aten._weight_int4pack_mm.default(buf729, arg410_1, 128, arg411_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg410_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg411_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf731 = buf730 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf730 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_50], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf733 = torch.ops.aten._weight_int4pack_mm.default(buf732, arg413_1, 128, arg414_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg413_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg414_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf734 = buf733 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf733 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf735 = reinterpret_tensor(buf728, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf728 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_239], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf735, arg409_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg409_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf736 = reinterpret_tensor(buf731, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf731 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_239], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf736, arg412_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg412_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf737 = reinterpret_tensor(buf734, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf734 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_239], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf737, arg415_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg415_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf738 = buf681; del buf681 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf795 = buf624; del buf624 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_239, hidden_states_258], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf738, buf795, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_239], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf739 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf735, buf736, buf737, reinterpret_tensor(buf738, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf735 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf736 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf737 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf740 = buf739[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf739 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf744 = buf726; del buf726 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_242], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf740, buf744, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_242], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf745 = torch.ops.aten._weight_int4pack_mm.default(buf744, arg416_1, 128, arg417_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg416_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg417_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf746 = buf745 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf745 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf750 = reinterpret_tensor(buf740, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf740 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_102, hidden_states_244, hidden_states_245, mul_54, norm_hidden_states_50, norm_hidden_states_51], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf746, arg418_1, buf725, arg394_1, buf9, arg14_1, buf750, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf751 = buf744; del buf744 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_246], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf750, buf751, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_246], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf752 = torch.ops.aten._weight_int4pack_mm.default(buf751, arg419_1, 128, arg420_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg419_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg420_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf753 = buf752 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf752 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf754 = buf697; del buf697 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_249], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf753, arg421_1, buf754, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg421_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf753 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_249], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf755 = torch.ops.aten._weight_int4pack_mm.default(buf754, arg422_1, 128, arg423_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg422_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg423_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf756 = buf755 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf755 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf757 = reinterpret_tensor(buf756, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf756 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf761 = buf750; del buf750 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_106, ff_output_12, hidden_states_244, hidden_states_245, hidden_states_250, mul_56, norm_hidden_states_52, norm_hidden_states_53], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf757, arg394_1, buf9, arg14_1, arg424_1, buf746, arg418_1, buf725, arg425_1, buf761, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg394_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg418_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg424_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf725 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf746 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf762 = buf751; del buf751 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf765 = buf708; del buf708 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf768 = buf705; del buf705 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_52, query_52, value_52], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf761, buf762, buf765, buf768, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf761 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_52], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf763 = torch.ops.aten._weight_int4pack_mm.default(buf762, arg426_1, 128, arg427_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg426_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg427_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf764 = buf763 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf763 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_52], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf766 = torch.ops.aten._weight_int4pack_mm.default(buf765, arg429_1, 128, arg430_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg429_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg430_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf767 = buf766 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf766 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_52], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf769 = torch.ops.aten._weight_int4pack_mm.default(buf768, arg432_1, 128, arg433_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg432_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg433_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf770 = buf769 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf769 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf771 = reinterpret_tensor(buf764, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf764 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_251], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf771, arg428_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg428_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf772 = reinterpret_tensor(buf767, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf767 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_251], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf772, arg431_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg431_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf773 = reinterpret_tensor(buf770, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf770 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_251], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf773, arg434_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg434_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_251], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf774 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf771, buf772, buf773, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf771 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf772 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf773 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf775 = buf774[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf774 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf779 = buf768; del buf768 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_254], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf775, buf779, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf775 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_254], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf780 = torch.ops.aten._weight_int4pack_mm.default(buf779, arg435_1, 128, arg436_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg435_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg436_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf781 = buf780 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf780 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf782 = reinterpret_tensor(buf781, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf781 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_13, hidden_states_256, hidden_states_257], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf782, arg425_1, buf9, arg14_1, arg437_1, buf757, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg437_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf757 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf783 = buf779; del buf779 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_54], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf782, buf783, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_54], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf784 = torch.ops.aten._weight_int4pack_mm.default(buf783, arg438_1, 128, arg439_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg438_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg439_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf785 = buf784 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf784 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_54], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf787 = torch.ops.aten._weight_int4pack_mm.default(buf786, arg441_1, 128, arg442_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg441_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg442_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf788 = buf787 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf787 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_54], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf790 = torch.ops.aten._weight_int4pack_mm.default(buf789, arg444_1, 128, arg445_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg444_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg445_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf791 = buf790 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf790 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf792 = reinterpret_tensor(buf785, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf785 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_258], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf792, arg440_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg440_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf793 = reinterpret_tensor(buf788, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf788 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_258], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf793, arg443_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg443_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf794 = reinterpret_tensor(buf791, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf791 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_258], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf794, arg446_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg446_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_258], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf796 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf792, buf793, buf794, reinterpret_tensor(buf795, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf792 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf793 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf794 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf797 = buf796[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf796 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf801 = buf783; del buf783 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_261], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf797, buf801, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_261], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf802 = torch.ops.aten._weight_int4pack_mm.default(buf801, arg447_1, 128, arg448_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg447_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg448_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf803 = buf802 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf802 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf807 = reinterpret_tensor(buf797, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf797 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_110, hidden_states_263, hidden_states_264, mul_58, norm_hidden_states_54, norm_hidden_states_55], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf803, arg449_1, buf782, arg425_1, buf9, arg14_1, buf807, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf808 = buf801; del buf801 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_265], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf807, buf808, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_265], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf809 = torch.ops.aten._weight_int4pack_mm.default(buf808, arg450_1, 128, arg451_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg450_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg451_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf810 = buf809 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf809 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf811 = buf754; del buf754 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_268], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf810, arg452_1, buf811, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg452_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf810 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_268], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf812 = torch.ops.aten._weight_int4pack_mm.default(buf811, arg453_1, 128, arg454_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg453_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg454_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf813 = buf812 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf812 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf814 = reinterpret_tensor(buf813, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf813 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf818 = buf807; del buf807 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_114, ff_output_13, hidden_states_263, hidden_states_264, hidden_states_269, mul_60, norm_hidden_states_56, norm_hidden_states_57], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf814, arg425_1, buf9, arg14_1, arg455_1, buf803, arg449_1, buf782, arg456_1, buf818, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg425_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg449_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg455_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf782 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf803 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf819 = buf808; del buf808 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf822 = buf765; del buf765 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf825 = buf762; del buf762 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_56, query_56, value_56], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf818, buf819, buf822, buf825, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf818 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_56], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf820 = torch.ops.aten._weight_int4pack_mm.default(buf819, arg457_1, 128, arg458_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg457_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg458_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf821 = buf820 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf820 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_56], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf823 = torch.ops.aten._weight_int4pack_mm.default(buf822, arg460_1, 128, arg461_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg460_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg461_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf824 = buf823 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf823 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_56], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf826 = torch.ops.aten._weight_int4pack_mm.default(buf825, arg463_1, 128, arg464_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg463_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg464_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf827 = buf826 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf826 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf828 = reinterpret_tensor(buf821, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf821 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_270], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf828, arg459_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg459_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf829 = reinterpret_tensor(buf824, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf824 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_270], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf829, arg462_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg462_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf830 = reinterpret_tensor(buf827, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf827 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_270], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf830, arg465_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg465_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_270], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf831 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf828, buf829, buf830, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf828 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf829 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf830 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf832 = buf831[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf831 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf836 = buf825; del buf825 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_273], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf832, buf836, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf832 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_273], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf837 = torch.ops.aten._weight_int4pack_mm.default(buf836, arg466_1, 128, arg467_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg466_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg467_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf838 = buf837 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf837 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf839 = reinterpret_tensor(buf838, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf838 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_14, hidden_states_275, hidden_states_276], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf839, arg456_1, buf9, arg14_1, arg468_1, buf814, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg468_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf814 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf840 = buf836; del buf836 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_58], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf839, buf840, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_58], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf841 = torch.ops.aten._weight_int4pack_mm.default(buf840, arg469_1, 128, arg470_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg469_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg470_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf842 = buf841 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf841 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf843 = buf789; del buf789 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf846 = buf786; del buf786 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf900 = buf732; del buf732 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf903 = buf729; del buf729 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_58, key_62, value_58, value_62], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf843, buf846, buf900, buf903, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_58], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf844 = torch.ops.aten._weight_int4pack_mm.default(buf843, arg472_1, 128, arg473_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg472_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg473_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf845 = buf844 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf844 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_58], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf847 = torch.ops.aten._weight_int4pack_mm.default(buf846, arg475_1, 128, arg476_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg475_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg476_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf848 = buf847 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf847 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf849 = reinterpret_tensor(buf842, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf842 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_277], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf849, arg471_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg471_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf850 = reinterpret_tensor(buf845, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf845 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_277], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf850, arg474_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg474_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf851 = reinterpret_tensor(buf848, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf848 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_277], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf851, arg477_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg477_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf852 = buf795; del buf795 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf909 = buf738; del buf738 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_277, hidden_states_296], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf852, buf909, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_277], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf853 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf849, buf850, buf851, reinterpret_tensor(buf852, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf849 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf850 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf851 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf854 = buf853[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf853 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf858 = buf840; del buf840 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_280], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf854, buf858, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_280], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf859 = torch.ops.aten._weight_int4pack_mm.default(buf858, arg478_1, 128, arg479_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg478_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg479_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf860 = buf859 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf859 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf864 = reinterpret_tensor(buf854, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf854 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_118, hidden_states_282, hidden_states_283, mul_62, norm_hidden_states_58, norm_hidden_states_59], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf860, arg480_1, buf839, arg456_1, buf9, arg14_1, buf864, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf865 = buf858; del buf858 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_284], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf864, buf865, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_284], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf866 = torch.ops.aten._weight_int4pack_mm.default(buf865, arg481_1, 128, arg482_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg481_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg482_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf867 = buf866 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf866 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf868 = buf811; del buf811 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_287], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf867, arg483_1, buf868, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg483_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf867 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_287], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf869 = torch.ops.aten._weight_int4pack_mm.default(buf868, arg484_1, 128, arg485_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg484_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg485_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf870 = buf869 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf869 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf871 = reinterpret_tensor(buf870, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf870 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf875 = buf864; del buf864 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_122, ff_output_14, hidden_states_282, hidden_states_283, hidden_states_288, mul_64, norm_hidden_states_60, norm_hidden_states_61], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf871, arg456_1, buf9, arg14_1, arg486_1, buf860, arg480_1, buf839, arg487_1, buf875, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg456_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg480_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg486_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf839 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf860 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf876 = buf865; del buf865 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf879 = buf822; del buf822 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf882 = buf819; del buf819 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_60, query_60, value_60], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf875, buf876, buf879, buf882, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf875 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_60], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf877 = torch.ops.aten._weight_int4pack_mm.default(buf876, arg488_1, 128, arg489_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg488_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg489_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf878 = buf877 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf877 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_60], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf880 = torch.ops.aten._weight_int4pack_mm.default(buf879, arg491_1, 128, arg492_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg491_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg492_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf881 = buf880 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf880 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_60], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf883 = torch.ops.aten._weight_int4pack_mm.default(buf882, arg494_1, 128, arg495_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg494_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg495_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf884 = buf883 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf883 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf885 = reinterpret_tensor(buf878, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf878 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_289], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf885, arg490_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg490_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf886 = reinterpret_tensor(buf881, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf881 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_289], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf886, arg493_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg493_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf887 = reinterpret_tensor(buf884, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf884 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_289], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf887, arg496_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg496_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_289], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf888 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf885, buf886, buf887, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf885 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf886 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf887 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf889 = buf888[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf888 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf893 = buf882; del buf882 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_292], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf889, buf893, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf889 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_292], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf894 = torch.ops.aten._weight_int4pack_mm.default(buf893, arg497_1, 128, arg498_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg497_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg498_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf895 = buf894 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf894 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf896 = reinterpret_tensor(buf895, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf895 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_15, hidden_states_294, hidden_states_295], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf896, arg487_1, buf9, arg14_1, arg499_1, buf871, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg499_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf871 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf897 = buf893; del buf893 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_62], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf896, buf897, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_62], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf898 = torch.ops.aten._weight_int4pack_mm.default(buf897, arg500_1, 128, arg501_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg500_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg501_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf899 = buf898 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf898 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_62], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf901 = torch.ops.aten._weight_int4pack_mm.default(buf900, arg503_1, 128, arg504_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg503_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg504_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf902 = buf901 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf901 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_62], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf904 = torch.ops.aten._weight_int4pack_mm.default(buf903, arg506_1, 128, arg507_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg506_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg507_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf905 = buf904 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf904 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf906 = reinterpret_tensor(buf899, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf899 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_296], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf906, arg502_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg502_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf907 = reinterpret_tensor(buf902, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf902 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_296], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf907, arg505_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg505_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf908 = reinterpret_tensor(buf905, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf905 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_296], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf908, arg508_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg508_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_296], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf910 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf906, buf907, buf908, reinterpret_tensor(buf909, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf906 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf907 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf908 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf911 = buf910[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf910 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf915 = buf897; del buf897 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_299], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf911, buf915, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_299], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf916 = torch.ops.aten._weight_int4pack_mm.default(buf915, arg509_1, 128, arg510_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg509_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg510_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf917 = buf916 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf916 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf921 = reinterpret_tensor(buf911, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf911 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_126, hidden_states_301, hidden_states_302, mul_66, norm_hidden_states_62, norm_hidden_states_63], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf917, arg511_1, buf896, arg487_1, buf9, arg14_1, buf921, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf922 = buf915; del buf915 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_303], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf921, buf922, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_303], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf923 = torch.ops.aten._weight_int4pack_mm.default(buf922, arg512_1, 128, arg513_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg512_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg513_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf924 = buf923 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf923 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf925 = buf868; del buf868 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_306], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf924, arg514_1, buf925, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg514_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf924 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_306], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf926 = torch.ops.aten._weight_int4pack_mm.default(buf925, arg515_1, 128, arg516_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg515_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg516_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf927 = buf926 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf926 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf928 = reinterpret_tensor(buf927, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf927 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf932 = buf921; del buf921 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_130, ff_output_15, hidden_states_301, hidden_states_302, hidden_states_307, mul_68, norm_hidden_states_64, norm_hidden_states_65], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf928, arg487_1, buf9, arg14_1, arg517_1, buf917, arg511_1, buf896, arg518_1, buf932, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg487_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg511_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg517_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf896 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf917 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf933 = buf922; del buf922 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf936 = buf879; del buf879 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf939 = buf876; del buf876 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_64, query_64, value_64], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf932, buf933, buf936, buf939, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf932 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_64], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf934 = torch.ops.aten._weight_int4pack_mm.default(buf933, arg519_1, 128, arg520_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg519_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg520_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf935 = buf934 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf934 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_64], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf937 = torch.ops.aten._weight_int4pack_mm.default(buf936, arg522_1, 128, arg523_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg522_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg523_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf938 = buf937 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf937 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_64], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf940 = torch.ops.aten._weight_int4pack_mm.default(buf939, arg525_1, 128, arg526_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg525_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg526_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf941 = buf940 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf940 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf942 = reinterpret_tensor(buf935, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf935 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_308], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf942, arg521_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg521_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf943 = reinterpret_tensor(buf938, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf938 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_308], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf943, arg524_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg524_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf944 = reinterpret_tensor(buf941, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf941 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_308], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf944, arg527_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg527_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_308], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf945 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf942, buf943, buf944, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf942 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf943 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf944 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf946 = buf945[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf945 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf950 = buf939; del buf939 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_311], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf946, buf950, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf946 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_311], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf951 = torch.ops.aten._weight_int4pack_mm.default(buf950, arg528_1, 128, arg529_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg528_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg529_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf952 = buf951 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf951 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf953 = reinterpret_tensor(buf952, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf952 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_16, hidden_states_313, hidden_states_314], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf953, arg518_1, buf9, arg14_1, arg530_1, buf928, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg530_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf928 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf954 = buf950; del buf950 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_66], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf953, buf954, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_66], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf955 = torch.ops.aten._weight_int4pack_mm.default(buf954, arg531_1, 128, arg532_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg531_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg532_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf956 = buf955 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf955 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf957 = buf903; del buf903 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf960 = buf900; del buf900 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1014 = buf846; del buf846 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1017 = buf843; del buf843 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_66, key_70, value_66, value_70], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf957, buf960, buf1014, buf1017, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_66], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf958 = torch.ops.aten._weight_int4pack_mm.default(buf957, arg534_1, 128, arg535_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg534_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg535_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf959 = buf958 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf958 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_66], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf961 = torch.ops.aten._weight_int4pack_mm.default(buf960, arg537_1, 128, arg538_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg537_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg538_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf962 = buf961 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf961 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf963 = reinterpret_tensor(buf956, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf956 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_315], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf963, arg533_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg533_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf964 = reinterpret_tensor(buf959, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf959 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_315], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf964, arg536_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg536_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf965 = reinterpret_tensor(buf962, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf962 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_315], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf965, arg539_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg539_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf966 = buf909; del buf909 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1023 = buf852; del buf852 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_315, hidden_states_334], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf966, buf1023, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_315], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf967 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf963, buf964, buf965, reinterpret_tensor(buf966, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf963 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf964 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf965 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf968 = buf967[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf967 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf972 = buf954; del buf954 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_318], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf968, buf972, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_318], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf973 = torch.ops.aten._weight_int4pack_mm.default(buf972, arg540_1, 128, arg541_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg540_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg541_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf974 = buf973 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf973 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf978 = reinterpret_tensor(buf968, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf968 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_134, hidden_states_320, hidden_states_321, mul_70, norm_hidden_states_66, norm_hidden_states_67], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf974, arg542_1, buf953, arg518_1, buf9, arg14_1, buf978, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf979 = buf972; del buf972 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_322], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf978, buf979, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_322], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf980 = torch.ops.aten._weight_int4pack_mm.default(buf979, arg543_1, 128, arg544_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg543_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg544_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf981 = buf980 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf980 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf982 = buf925; del buf925 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_325], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf981, arg545_1, buf982, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg545_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf981 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_325], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf983 = torch.ops.aten._weight_int4pack_mm.default(buf982, arg546_1, 128, arg547_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg546_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg547_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf984 = buf983 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf983 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf985 = reinterpret_tensor(buf984, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf984 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf989 = buf978; del buf978 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_138, ff_output_16, hidden_states_320, hidden_states_321, hidden_states_326, mul_72, norm_hidden_states_68, norm_hidden_states_69], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf985, arg518_1, buf9, arg14_1, arg548_1, buf974, arg542_1, buf953, arg549_1, buf989, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg518_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg542_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg548_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf953 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf974 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf990 = buf979; del buf979 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf993 = buf936; del buf936 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf996 = buf933; del buf933 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_68, query_68, value_68], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf989, buf990, buf993, buf996, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf989 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_68], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf991 = torch.ops.aten._weight_int4pack_mm.default(buf990, arg550_1, 128, arg551_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg550_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg551_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf992 = buf991 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf991 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_68], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf994 = torch.ops.aten._weight_int4pack_mm.default(buf993, arg553_1, 128, arg554_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg553_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg554_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf995 = buf994 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf994 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_68], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf997 = torch.ops.aten._weight_int4pack_mm.default(buf996, arg556_1, 128, arg557_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg556_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg557_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf998 = buf997 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf997 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf999 = reinterpret_tensor(buf992, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf992 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_327], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf999, arg552_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg552_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1000 = reinterpret_tensor(buf995, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf995 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_327], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1000, arg555_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg555_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1001 = reinterpret_tensor(buf998, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf998 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_327], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1001, arg558_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg558_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_327], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1002 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf999, buf1000, buf1001, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1000 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1001 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf999 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1003 = buf1002[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1002 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1007 = buf996; del buf996 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_330], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1003, buf1007, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1003 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_330], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1008 = torch.ops.aten._weight_int4pack_mm.default(buf1007, arg559_1, 128, arg560_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg559_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg560_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1009 = buf1008 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1008 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1010 = reinterpret_tensor(buf1009, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1009 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_17, hidden_states_332, hidden_states_333], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1010, arg549_1, buf9, arg14_1, arg561_1, buf985, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg561_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf985 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1011 = buf1007; del buf1007 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_70], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1010, buf1011, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_70], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1012 = torch.ops.aten._weight_int4pack_mm.default(buf1011, arg562_1, 128, arg563_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg562_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg563_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1013 = buf1012 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1012 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_70], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1015 = torch.ops.aten._weight_int4pack_mm.default(buf1014, arg565_1, 128, arg566_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg565_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg566_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1016 = buf1015 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1015 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_70], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1018 = torch.ops.aten._weight_int4pack_mm.default(buf1017, arg568_1, 128, arg569_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg568_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg569_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1019 = buf1018 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1018 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1020 = reinterpret_tensor(buf1013, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1013 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_334], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1020, arg564_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg564_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1021 = reinterpret_tensor(buf1016, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1016 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_334], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1021, arg567_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg567_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1022 = reinterpret_tensor(buf1019, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1019 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_334], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1022, arg570_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg570_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_334], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1024 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1020, buf1021, buf1022, reinterpret_tensor(buf1023, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1020 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1021 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1022 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1025 = buf1024[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1024 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1029 = buf1011; del buf1011 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_337], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1025, buf1029, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_337], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1030 = torch.ops.aten._weight_int4pack_mm.default(buf1029, arg571_1, 128, arg572_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg571_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg572_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1031 = buf1030 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1030 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1035 = reinterpret_tensor(buf1025, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1025 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_142, hidden_states_339, hidden_states_340, mul_74, norm_hidden_states_70, norm_hidden_states_71], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1031, arg573_1, buf1010, arg549_1, buf9, arg14_1, buf1035, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1036 = buf1029; del buf1029 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_341], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1035, buf1036, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_341], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1037 = torch.ops.aten._weight_int4pack_mm.default(buf1036, arg574_1, 128, arg575_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg574_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg575_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1038 = buf1037 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1037 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1039 = buf982; del buf982 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_344], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1038, arg576_1, buf1039, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg576_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1038 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_344], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1040 = torch.ops.aten._weight_int4pack_mm.default(buf1039, arg577_1, 128, arg578_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg577_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg578_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1041 = buf1040 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1040 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1042 = reinterpret_tensor(buf1041, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1041 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1046 = buf1035; del buf1035 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_146, ff_output_17, hidden_states_339, hidden_states_340, hidden_states_345, mul_76, norm_hidden_states_72, norm_hidden_states_73], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1042, arg549_1, buf9, arg14_1, arg579_1, buf1031, arg573_1, buf1010, arg580_1, buf1046, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg549_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg573_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg579_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1010 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1031 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1047 = buf1036; del buf1036 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1050 = buf993; del buf993 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1053 = buf990; del buf990 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_72, query_72, value_72], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1046, buf1047, buf1050, buf1053, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1046 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_72], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1048 = torch.ops.aten._weight_int4pack_mm.default(buf1047, arg581_1, 128, arg582_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg581_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg582_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1049 = buf1048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1048 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_72], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1051 = torch.ops.aten._weight_int4pack_mm.default(buf1050, arg584_1, 128, arg585_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg584_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg585_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1052 = buf1051 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1051 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_72], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1054 = torch.ops.aten._weight_int4pack_mm.default(buf1053, arg587_1, 128, arg588_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg587_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg588_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1055 = buf1054 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1054 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1056 = reinterpret_tensor(buf1049, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1049 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_346], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1056, arg583_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg583_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1057 = reinterpret_tensor(buf1052, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1052 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_346], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1057, arg586_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg586_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1058 = reinterpret_tensor(buf1055, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1055 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_346], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1058, arg589_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg589_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_346], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1059 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1056, buf1057, buf1058, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1056 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1057 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1058 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1060 = buf1059[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1059 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1064 = buf1053; del buf1053 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_349], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1060, buf1064, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1060 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_349], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1065 = torch.ops.aten._weight_int4pack_mm.default(buf1064, arg590_1, 128, arg591_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg590_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg591_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1066 = buf1065 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1065 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1067 = reinterpret_tensor(buf1066, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1066 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_18, hidden_states_351, hidden_states_352], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1067, arg580_1, buf9, arg14_1, arg592_1, buf1042, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg592_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1042 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1068 = buf1064; del buf1064 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_74], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1067, buf1068, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_74], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1069 = torch.ops.aten._weight_int4pack_mm.default(buf1068, arg593_1, 128, arg594_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg593_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg594_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1070 = buf1069 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1069 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1071 = buf1017; del buf1017 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1074 = buf1014; del buf1014 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1128 = buf960; del buf960 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1131 = buf957; del buf957 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_74, key_78, value_74, value_78], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1071, buf1074, buf1128, buf1131, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_74], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1072 = torch.ops.aten._weight_int4pack_mm.default(buf1071, arg596_1, 128, arg597_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg596_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg597_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1073 = buf1072 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1072 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_74], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1075 = torch.ops.aten._weight_int4pack_mm.default(buf1074, arg599_1, 128, arg600_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg599_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg600_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1076 = buf1075 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1075 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1077 = reinterpret_tensor(buf1070, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1070 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_353], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1077, arg595_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg595_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1078 = reinterpret_tensor(buf1073, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1073 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_353], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1078, arg598_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg598_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1079 = reinterpret_tensor(buf1076, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1076 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_353], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1079, arg601_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg601_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1080 = buf1023; del buf1023 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1137 = buf966; del buf966 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_353, hidden_states_372], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1080, buf1137, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_353], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1081 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1077, buf1078, buf1079, reinterpret_tensor(buf1080, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1077 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1078 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1079 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1082 = buf1081[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1081 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1086 = buf1068; del buf1068 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_356], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1082, buf1086, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_356], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1087 = torch.ops.aten._weight_int4pack_mm.default(buf1086, arg602_1, 128, arg603_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg602_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg603_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1088 = buf1087 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1087 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1092 = reinterpret_tensor(buf1082, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1082 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_150, hidden_states_358, hidden_states_359, mul_78, norm_hidden_states_74, norm_hidden_states_75], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1088, arg604_1, buf1067, arg580_1, buf9, arg14_1, buf1092, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1093 = buf1086; del buf1086 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_360], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1092, buf1093, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_360], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1094 = torch.ops.aten._weight_int4pack_mm.default(buf1093, arg605_1, 128, arg606_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg605_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg606_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1095 = buf1094 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1094 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1096 = buf1039; del buf1039 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_363], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1095, arg607_1, buf1096, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg607_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1095 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_363], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1097 = torch.ops.aten._weight_int4pack_mm.default(buf1096, arg608_1, 128, arg609_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg608_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg609_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1098 = buf1097 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1097 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1099 = reinterpret_tensor(buf1098, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1098 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1103 = buf1092; del buf1092 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_154, ff_output_18, hidden_states_358, hidden_states_359, hidden_states_364, mul_80, norm_hidden_states_76, norm_hidden_states_77], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1099, arg580_1, buf9, arg14_1, arg610_1, buf1088, arg604_1, buf1067, arg611_1, buf1103, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg580_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg604_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg610_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1067 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1088 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1104 = buf1093; del buf1093 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1107 = buf1050; del buf1050 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1110 = buf1047; del buf1047 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_76, query_76, value_76], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1103, buf1104, buf1107, buf1110, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1103 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_76], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1105 = torch.ops.aten._weight_int4pack_mm.default(buf1104, arg612_1, 128, arg613_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg612_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg613_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1106 = buf1105 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1105 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_76], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1108 = torch.ops.aten._weight_int4pack_mm.default(buf1107, arg615_1, 128, arg616_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg615_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg616_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1109 = buf1108 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1108 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_76], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1111 = torch.ops.aten._weight_int4pack_mm.default(buf1110, arg618_1, 128, arg619_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg618_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg619_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1112 = buf1111 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1111 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1113 = reinterpret_tensor(buf1106, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1106 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_365], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1113, arg614_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg614_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1114 = reinterpret_tensor(buf1109, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1109 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_365], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1114, arg617_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg617_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1115 = reinterpret_tensor(buf1112, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1112 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_365], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1115, arg620_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg620_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_365], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1116 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1113, buf1114, buf1115, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1113 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1114 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1115 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1117 = buf1116[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1116 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1121 = buf1110; del buf1110 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_368], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1117, buf1121, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1117 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_368], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1122 = torch.ops.aten._weight_int4pack_mm.default(buf1121, arg621_1, 128, arg622_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg621_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg622_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1123 = buf1122 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1122 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1124 = reinterpret_tensor(buf1123, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1123 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_19, hidden_states_370, hidden_states_371], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1124, arg611_1, buf9, arg14_1, arg623_1, buf1099, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg623_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1099 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1125 = buf1121; del buf1121 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_78], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1124, buf1125, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_78], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1126 = torch.ops.aten._weight_int4pack_mm.default(buf1125, arg624_1, 128, arg625_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg624_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg625_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1127 = buf1126 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1126 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_78], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1129 = torch.ops.aten._weight_int4pack_mm.default(buf1128, arg627_1, 128, arg628_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg627_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg628_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1130 = buf1129 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1129 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_78], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1132 = torch.ops.aten._weight_int4pack_mm.default(buf1131, arg630_1, 128, arg631_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg630_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg631_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1133 = buf1132 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1132 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1134 = reinterpret_tensor(buf1127, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1127 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_372], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1134, arg626_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg626_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1135 = reinterpret_tensor(buf1130, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1130 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_372], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1135, arg629_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg629_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1136 = reinterpret_tensor(buf1133, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1133 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_372], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1136, arg632_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg632_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_372], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1138 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1134, buf1135, buf1136, reinterpret_tensor(buf1137, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1134 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1135 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1136 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1139 = buf1138[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1138 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1143 = buf1125; del buf1125 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_375], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1139, buf1143, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_375], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1144 = torch.ops.aten._weight_int4pack_mm.default(buf1143, arg633_1, 128, arg634_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg633_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg634_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1145 = buf1144 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1144 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1149 = reinterpret_tensor(buf1139, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1139 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_158, hidden_states_377, hidden_states_378, mul_82, norm_hidden_states_78, norm_hidden_states_79], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1145, arg635_1, buf1124, arg611_1, buf9, arg14_1, buf1149, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1150 = buf1143; del buf1143 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_379], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1149, buf1150, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_379], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1151 = torch.ops.aten._weight_int4pack_mm.default(buf1150, arg636_1, 128, arg637_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg636_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg637_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1152 = buf1151 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1151 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1153 = buf1096; del buf1096 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_382], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1152, arg638_1, buf1153, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg638_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1152 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_382], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1154 = torch.ops.aten._weight_int4pack_mm.default(buf1153, arg639_1, 128, arg640_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg639_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg640_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1155 = buf1154 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1154 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1156 = reinterpret_tensor(buf1155, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1155 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1160 = buf1149; del buf1149 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_162, ff_output_19, hidden_states_377, hidden_states_378, hidden_states_383, mul_84, norm_hidden_states_80, norm_hidden_states_81], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1156, arg611_1, buf9, arg14_1, arg641_1, buf1145, arg635_1, buf1124, arg642_1, buf1160, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg611_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg635_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg641_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1124 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1145 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1161 = buf1150; del buf1150 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1164 = buf1107; del buf1107 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1167 = buf1104; del buf1104 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_80, query_80, value_80], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1160, buf1161, buf1164, buf1167, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1160 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_80], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1162 = torch.ops.aten._weight_int4pack_mm.default(buf1161, arg643_1, 128, arg644_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg643_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg644_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1163 = buf1162 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1162 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_80], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1165 = torch.ops.aten._weight_int4pack_mm.default(buf1164, arg646_1, 128, arg647_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg646_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg647_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1166 = buf1165 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1165 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_80], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1168 = torch.ops.aten._weight_int4pack_mm.default(buf1167, arg649_1, 128, arg650_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg649_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg650_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1169 = buf1168 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1168 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1170 = reinterpret_tensor(buf1163, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1163 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_384], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1170, arg645_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg645_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1171 = reinterpret_tensor(buf1166, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1166 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_384], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1171, arg648_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg648_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1172 = reinterpret_tensor(buf1169, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1169 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_384], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1172, arg651_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg651_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_384], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1173 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1170, buf1171, buf1172, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1170 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1171 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1172 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1174 = buf1173[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1173 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1178 = buf1167; del buf1167 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_387], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1174, buf1178, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1174 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_387], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1179 = torch.ops.aten._weight_int4pack_mm.default(buf1178, arg652_1, 128, arg653_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg652_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg653_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1180 = buf1179 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1179 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1181 = reinterpret_tensor(buf1180, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1180 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_20, hidden_states_389, hidden_states_390], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1181, arg642_1, buf9, arg14_1, arg654_1, buf1156, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg654_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1156 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1182 = buf1178; del buf1178 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_82], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1181, buf1182, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_82], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1183 = torch.ops.aten._weight_int4pack_mm.default(buf1182, arg655_1, 128, arg656_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg655_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg656_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1184 = buf1183 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1183 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1185 = buf1131; del buf1131 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1188 = buf1128; del buf1128 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1242 = buf1074; del buf1074 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1245 = buf1071; del buf1071 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_82, key_86, value_82, value_86], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1185, buf1188, buf1242, buf1245, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_82], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1186 = torch.ops.aten._weight_int4pack_mm.default(buf1185, arg658_1, 128, arg659_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg658_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg659_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1187 = buf1186 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1186 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_82], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1189 = torch.ops.aten._weight_int4pack_mm.default(buf1188, arg661_1, 128, arg662_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg661_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg662_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1190 = buf1189 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1189 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1191 = reinterpret_tensor(buf1184, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1184 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_391], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1191, arg657_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg657_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1192 = reinterpret_tensor(buf1187, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1187 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_391], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1192, arg660_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg660_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1193 = reinterpret_tensor(buf1190, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1190 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_391], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1193, arg663_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg663_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1194 = buf1137; del buf1137 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1251 = buf1080; del buf1080 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_391, hidden_states_410], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1194, buf1251, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_391], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1195 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1191, buf1192, buf1193, reinterpret_tensor(buf1194, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1191 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1192 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1193 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1196 = buf1195[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1195 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1200 = buf1182; del buf1182 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_394], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1196, buf1200, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_394], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1201 = torch.ops.aten._weight_int4pack_mm.default(buf1200, arg664_1, 128, arg665_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg664_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg665_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1202 = buf1201 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1201 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1206 = reinterpret_tensor(buf1196, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1196 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_166, hidden_states_396, hidden_states_397, mul_86, norm_hidden_states_82, norm_hidden_states_83], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1202, arg666_1, buf1181, arg642_1, buf9, arg14_1, buf1206, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1207 = buf1200; del buf1200 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_398], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1206, buf1207, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_398], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1208 = torch.ops.aten._weight_int4pack_mm.default(buf1207, arg667_1, 128, arg668_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg667_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg668_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1209 = buf1208 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1208 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1210 = buf1153; del buf1153 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_401], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1209, arg669_1, buf1210, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg669_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1209 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_401], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1211 = torch.ops.aten._weight_int4pack_mm.default(buf1210, arg670_1, 128, arg671_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg670_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg671_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1212 = buf1211 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1211 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1213 = reinterpret_tensor(buf1212, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1212 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1217 = buf1206; del buf1206 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_170, ff_output_20, hidden_states_396, hidden_states_397, hidden_states_402, mul_88, norm_hidden_states_84, norm_hidden_states_85], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1213, arg642_1, buf9, arg14_1, arg672_1, buf1202, arg666_1, buf1181, arg673_1, buf1217, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg642_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg666_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg672_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1181 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1202 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1218 = buf1207; del buf1207 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1221 = buf1164; del buf1164 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1224 = buf1161; del buf1161 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_84, query_84, value_84], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1217, buf1218, buf1221, buf1224, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1217 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_84], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1219 = torch.ops.aten._weight_int4pack_mm.default(buf1218, arg674_1, 128, arg675_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg674_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg675_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1220 = buf1219 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1219 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_84], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1222 = torch.ops.aten._weight_int4pack_mm.default(buf1221, arg677_1, 128, arg678_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg677_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg678_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1223 = buf1222 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1222 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_84], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1225 = torch.ops.aten._weight_int4pack_mm.default(buf1224, arg680_1, 128, arg681_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg680_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg681_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1226 = buf1225 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1225 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1227 = reinterpret_tensor(buf1220, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1220 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_403], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1227, arg676_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg676_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1228 = reinterpret_tensor(buf1223, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1223 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_403], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1228, arg679_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg679_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1229 = reinterpret_tensor(buf1226, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1226 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_403], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1229, arg682_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg682_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_403], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1230 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1227, buf1228, buf1229, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1227 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1228 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1229 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1231 = buf1230[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1230 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1235 = buf1224; del buf1224 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_406], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1231, buf1235, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1231 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_406], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1236 = torch.ops.aten._weight_int4pack_mm.default(buf1235, arg683_1, 128, arg684_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg683_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg684_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1237 = buf1236 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1236 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1238 = reinterpret_tensor(buf1237, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1237 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_21, hidden_states_408, hidden_states_409], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1238, arg673_1, buf9, arg14_1, arg685_1, buf1213, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg685_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1213 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1239 = buf1235; del buf1235 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_86], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1238, buf1239, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_86], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1240 = torch.ops.aten._weight_int4pack_mm.default(buf1239, arg686_1, 128, arg687_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg686_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg687_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1241 = buf1240 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1240 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_86], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1243 = torch.ops.aten._weight_int4pack_mm.default(buf1242, arg689_1, 128, arg690_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg689_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg690_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1244 = buf1243 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1243 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_86], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1246 = torch.ops.aten._weight_int4pack_mm.default(buf1245, arg692_1, 128, arg693_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg692_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg693_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1247 = buf1246 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1246 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1248 = reinterpret_tensor(buf1241, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1241 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_410], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1248, arg688_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg688_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1249 = reinterpret_tensor(buf1244, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1244 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_410], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1249, arg691_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg691_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1250 = reinterpret_tensor(buf1247, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1247 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_410], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1250, arg694_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg694_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_410], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1252 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1248, buf1249, buf1250, reinterpret_tensor(buf1251, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1248 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1249 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1250 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1253 = buf1252[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1252 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1257 = buf1239; del buf1239 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_413], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1253, buf1257, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_413], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1258 = torch.ops.aten._weight_int4pack_mm.default(buf1257, arg695_1, 128, arg696_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg695_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg696_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1259 = buf1258 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1258 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1263 = reinterpret_tensor(buf1253, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1253 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_174, hidden_states_415, hidden_states_416, mul_90, norm_hidden_states_86, norm_hidden_states_87], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1259, arg697_1, buf1238, arg673_1, buf9, arg14_1, buf1263, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1264 = buf1257; del buf1257 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_417], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1263, buf1264, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_417], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1265 = torch.ops.aten._weight_int4pack_mm.default(buf1264, arg698_1, 128, arg699_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg698_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg699_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1266 = buf1265 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1265 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1267 = buf1210; del buf1210 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_420], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1266, arg700_1, buf1267, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg700_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1266 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_420], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1268 = torch.ops.aten._weight_int4pack_mm.default(buf1267, arg701_1, 128, arg702_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg701_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg702_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1269 = buf1268 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1268 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1270 = reinterpret_tensor(buf1269, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1269 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1274 = buf1263; del buf1263 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_178, ff_output_21, hidden_states_415, hidden_states_416, hidden_states_421, mul_92, norm_hidden_states_88, norm_hidden_states_89], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1270, arg673_1, buf9, arg14_1, arg703_1, buf1259, arg697_1, buf1238, arg704_1, buf1274, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg673_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg697_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg703_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1238 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1259 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1275 = buf1264; del buf1264 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1278 = buf1221; del buf1221 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1281 = buf1218; del buf1218 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_88, query_88, value_88], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1274, buf1275, buf1278, buf1281, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1274 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_88], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1276 = torch.ops.aten._weight_int4pack_mm.default(buf1275, arg705_1, 128, arg706_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg705_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg706_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1277 = buf1276 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1276 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_88], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1279 = torch.ops.aten._weight_int4pack_mm.default(buf1278, arg708_1, 128, arg709_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg708_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg709_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1280 = buf1279 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1279 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_88], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1282 = torch.ops.aten._weight_int4pack_mm.default(buf1281, arg711_1, 128, arg712_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg711_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg712_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1283 = buf1282 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1282 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1284 = reinterpret_tensor(buf1277, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1277 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_422], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1284, arg707_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg707_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1285 = reinterpret_tensor(buf1280, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1280 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_422], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1285, arg710_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg710_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1286 = reinterpret_tensor(buf1283, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1283 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_422], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1286, arg713_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg713_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_422], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1287 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1284, buf1285, buf1286, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1284 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1285 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1286 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1288 = buf1287[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1287 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1292 = buf1281; del buf1281 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_425], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1288, buf1292, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1288 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_425], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1293 = torch.ops.aten._weight_int4pack_mm.default(buf1292, arg714_1, 128, arg715_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg714_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg715_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1294 = buf1293 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1293 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1295 = reinterpret_tensor(buf1294, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1294 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_22, hidden_states_427, hidden_states_428], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1295, arg704_1, buf9, arg14_1, arg716_1, buf1270, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg716_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1270 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1296 = buf1292; del buf1292 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_90], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1295, buf1296, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_90], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1297 = torch.ops.aten._weight_int4pack_mm.default(buf1296, arg717_1, 128, arg718_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg717_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg718_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1298 = buf1297 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1297 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1299 = buf1245; del buf1245 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1302 = buf1242; del buf1242 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1356 = buf1188; del buf1188 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1359 = buf1185; del buf1185 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_90, key_94, value_90, value_94], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1299, buf1302, buf1356, buf1359, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_90], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1300 = torch.ops.aten._weight_int4pack_mm.default(buf1299, arg720_1, 128, arg721_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg720_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg721_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1301 = buf1300 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1300 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_90], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1303 = torch.ops.aten._weight_int4pack_mm.default(buf1302, arg723_1, 128, arg724_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg723_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg724_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1304 = buf1303 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1303 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1305 = reinterpret_tensor(buf1298, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1298 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_429], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1305, arg719_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg719_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1306 = reinterpret_tensor(buf1301, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1301 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_429], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1306, arg722_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg722_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1307 = reinterpret_tensor(buf1304, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1304 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_429], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1307, arg725_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg725_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1308 = buf1251; del buf1251 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1365 = buf1194; del buf1194 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_429, hidden_states_448], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1308, buf1365, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_429], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1309 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1305, buf1306, buf1307, reinterpret_tensor(buf1308, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1305 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1306 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1307 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1310 = buf1309[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1309 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1314 = buf1296; del buf1296 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_432], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1310, buf1314, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_432], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1315 = torch.ops.aten._weight_int4pack_mm.default(buf1314, arg726_1, 128, arg727_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg726_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg727_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1316 = buf1315 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1315 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1320 = reinterpret_tensor(buf1310, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1310 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_182, hidden_states_434, hidden_states_435, mul_94, norm_hidden_states_90, norm_hidden_states_91], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1316, arg728_1, buf1295, arg704_1, buf9, arg14_1, buf1320, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1321 = buf1314; del buf1314 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_436], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1320, buf1321, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_436], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1322 = torch.ops.aten._weight_int4pack_mm.default(buf1321, arg729_1, 128, arg730_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg729_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg730_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1323 = buf1322 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1322 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1324 = buf1267; del buf1267 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_439], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1323, arg731_1, buf1324, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg731_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1323 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_439], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1325 = torch.ops.aten._weight_int4pack_mm.default(buf1324, arg732_1, 128, arg733_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg732_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg733_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1326 = buf1325 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1325 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1327 = reinterpret_tensor(buf1326, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1326 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1331 = buf1320; del buf1320 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_186, ff_output_22, hidden_states_434, hidden_states_435, hidden_states_440, mul_96, norm_hidden_states_92, norm_hidden_states_93], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1327, arg704_1, buf9, arg14_1, arg734_1, buf1316, arg728_1, buf1295, arg735_1, buf1331, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg704_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg728_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg734_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1295 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1316 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1332 = buf1321; del buf1321 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1335 = buf1278; del buf1278 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1338 = buf1275; del buf1275 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_92, query_92, value_92], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1331, buf1332, buf1335, buf1338, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1331 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_92], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1333 = torch.ops.aten._weight_int4pack_mm.default(buf1332, arg736_1, 128, arg737_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg736_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg737_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1334 = buf1333 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1333 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_92], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1336 = torch.ops.aten._weight_int4pack_mm.default(buf1335, arg739_1, 128, arg740_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg739_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg740_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1337 = buf1336 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1336 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_92], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1339 = torch.ops.aten._weight_int4pack_mm.default(buf1338, arg742_1, 128, arg743_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg742_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg743_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1340 = buf1339 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1339 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1341 = reinterpret_tensor(buf1334, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1334 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_441], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1341, arg738_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg738_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1342 = reinterpret_tensor(buf1337, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1337 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_441], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1342, arg741_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg741_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1343 = reinterpret_tensor(buf1340, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1340 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_441], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1343, arg744_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg744_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_441], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1344 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1341, buf1342, buf1343, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1341 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1342 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1343 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1345 = buf1344[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1344 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1349 = buf1338; del buf1338 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_444], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1345, buf1349, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1345 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_444], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1350 = torch.ops.aten._weight_int4pack_mm.default(buf1349, arg745_1, 128, arg746_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg745_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg746_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1351 = buf1350 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1350 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1352 = reinterpret_tensor(buf1351, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1351 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_23, hidden_states_446, hidden_states_447], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1352, arg735_1, buf9, arg14_1, arg747_1, buf1327, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg747_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1327 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1353 = buf1349; del buf1349 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_94], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1352, buf1353, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_94], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1354 = torch.ops.aten._weight_int4pack_mm.default(buf1353, arg748_1, 128, arg749_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg748_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg749_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1355 = buf1354 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1354 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_94], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1357 = torch.ops.aten._weight_int4pack_mm.default(buf1356, arg751_1, 128, arg752_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg751_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg752_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1358 = buf1357 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1357 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_94], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1360 = torch.ops.aten._weight_int4pack_mm.default(buf1359, arg754_1, 128, arg755_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg754_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg755_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1361 = buf1360 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1360 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1362 = reinterpret_tensor(buf1355, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1355 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_448], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1362, arg750_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg750_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1363 = reinterpret_tensor(buf1358, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1358 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_448], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1363, arg753_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg753_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1364 = reinterpret_tensor(buf1361, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1361 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_448], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1364, arg756_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg756_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_448], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1366 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1362, buf1363, buf1364, reinterpret_tensor(buf1365, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1362 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1363 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1364 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1367 = buf1366[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1366 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1371 = buf1353; del buf1353 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_451], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1367, buf1371, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_451], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1372 = torch.ops.aten._weight_int4pack_mm.default(buf1371, arg757_1, 128, arg758_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg757_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg758_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1373 = buf1372 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1372 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1377 = reinterpret_tensor(buf1367, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1367 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_190, hidden_states_453, hidden_states_454, mul_98, norm_hidden_states_94, norm_hidden_states_95], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1373, arg759_1, buf1352, arg735_1, buf9, arg14_1, buf1377, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1378 = buf1371; del buf1371 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_455], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1377, buf1378, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_455], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1379 = torch.ops.aten._weight_int4pack_mm.default(buf1378, arg760_1, 128, arg761_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg760_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg761_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1380 = buf1379 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1379 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1381 = buf1324; del buf1324 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_458], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1380, arg762_1, buf1381, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg762_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1380 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_458], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1382 = torch.ops.aten._weight_int4pack_mm.default(buf1381, arg763_1, 128, arg764_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg763_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg764_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1383 = buf1382 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1382 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1384 = reinterpret_tensor(buf1383, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1383 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1388 = buf1377; del buf1377 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_194, ff_output_23, hidden_states_453, hidden_states_454, hidden_states_459, mul_100, norm_hidden_states_96, norm_hidden_states_97], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1384, arg735_1, buf9, arg14_1, arg765_1, buf1373, arg759_1, buf1352, arg766_1, buf1388, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg735_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg759_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg765_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1352 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1373 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1389 = buf1378; del buf1378 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1392 = buf1335; del buf1335 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1395 = buf1332; del buf1332 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_96, query_96, value_96], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1388, buf1389, buf1392, buf1395, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1388 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_96], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1390 = torch.ops.aten._weight_int4pack_mm.default(buf1389, arg767_1, 128, arg768_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg767_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg768_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1391 = buf1390 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1390 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_96], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1393 = torch.ops.aten._weight_int4pack_mm.default(buf1392, arg770_1, 128, arg771_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg770_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg771_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1394 = buf1393 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1393 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_96], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1396 = torch.ops.aten._weight_int4pack_mm.default(buf1395, arg773_1, 128, arg774_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg773_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg774_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1397 = buf1396 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1396 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1398 = reinterpret_tensor(buf1391, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1391 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_460], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1398, arg769_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg769_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1399 = reinterpret_tensor(buf1394, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1394 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_460], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1399, arg772_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg772_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1400 = reinterpret_tensor(buf1397, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1397 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_460], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1400, arg775_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg775_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_460], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1401 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1398, buf1399, buf1400, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1398 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1399 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1400 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1402 = buf1401[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1401 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1406 = buf1395; del buf1395 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_463], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1402, buf1406, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1402 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_463], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1407 = torch.ops.aten._weight_int4pack_mm.default(buf1406, arg776_1, 128, arg777_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg776_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg777_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1408 = buf1407 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1407 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1409 = reinterpret_tensor(buf1408, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1408 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_24, hidden_states_465, hidden_states_466], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1409, arg766_1, buf9, arg14_1, arg778_1, buf1384, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg778_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1384 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1410 = buf1406; del buf1406 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_98], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1409, buf1410, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_98], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1411 = torch.ops.aten._weight_int4pack_mm.default(buf1410, arg779_1, 128, arg780_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg779_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg780_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1412 = buf1411 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1411 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1413 = buf1359; del buf1359 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1416 = buf1356; del buf1356 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1470 = buf1302; del buf1302 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1473 = buf1299; del buf1299 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_102, key_98, value_102, value_98], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1413, buf1416, buf1470, buf1473, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_98], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1414 = torch.ops.aten._weight_int4pack_mm.default(buf1413, arg782_1, 128, arg783_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg782_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg783_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1415 = buf1414 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1414 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_98], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1417 = torch.ops.aten._weight_int4pack_mm.default(buf1416, arg785_1, 128, arg786_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg785_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg786_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1418 = buf1417 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1417 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1419 = reinterpret_tensor(buf1412, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1412 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_467], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1419, arg781_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg781_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1420 = reinterpret_tensor(buf1415, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1415 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_467], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1420, arg784_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg784_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1421 = reinterpret_tensor(buf1418, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1418 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_467], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1421, arg787_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg787_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1422 = buf1365; del buf1365 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1479 = buf1308; del buf1308 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_467, hidden_states_486], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1422, buf1479, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_467], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1423 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1419, buf1420, buf1421, reinterpret_tensor(buf1422, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1419 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1420 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1421 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1424 = buf1423[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1423 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1428 = buf1410; del buf1410 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_470], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1424, buf1428, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_470], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1429 = torch.ops.aten._weight_int4pack_mm.default(buf1428, arg788_1, 128, arg789_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg788_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg789_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1430 = buf1429 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1429 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1434 = reinterpret_tensor(buf1424, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1424 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_198, hidden_states_472, hidden_states_473, mul_102, norm_hidden_states_98, norm_hidden_states_99], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1430, arg790_1, buf1409, arg766_1, buf9, arg14_1, buf1434, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1435 = buf1428; del buf1428 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_474], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1434, buf1435, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_474], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1436 = torch.ops.aten._weight_int4pack_mm.default(buf1435, arg791_1, 128, arg792_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg791_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg792_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1437 = buf1436 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1436 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1438 = buf1381; del buf1381 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_477], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1437, arg793_1, buf1438, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg793_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1437 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_477], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1439 = torch.ops.aten._weight_int4pack_mm.default(buf1438, arg794_1, 128, arg795_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg794_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg795_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1440 = buf1439 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1439 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1441 = reinterpret_tensor(buf1440, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1440 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1445 = buf1434; del buf1434 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_202, ff_output_24, hidden_states_472, hidden_states_473, hidden_states_478, mul_104, norm_hidden_states_100, norm_hidden_states_101], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1441, arg766_1, buf9, arg14_1, arg796_1, buf1430, arg790_1, buf1409, arg797_1, buf1445, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg766_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg790_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg796_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1409 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1430 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1446 = buf1435; del buf1435 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1449 = buf1392; del buf1392 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1452 = buf1389; del buf1389 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_100, query_100, value_100], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1445, buf1446, buf1449, buf1452, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1445 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_100], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1447 = torch.ops.aten._weight_int4pack_mm.default(buf1446, arg798_1, 128, arg799_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg798_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg799_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1448 = buf1447 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1447 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_100], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1450 = torch.ops.aten._weight_int4pack_mm.default(buf1449, arg801_1, 128, arg802_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg801_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg802_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1451 = buf1450 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1450 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_100], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1453 = torch.ops.aten._weight_int4pack_mm.default(buf1452, arg804_1, 128, arg805_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg804_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg805_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1454 = buf1453 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1453 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1455 = reinterpret_tensor(buf1448, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1448 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_479], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1455, arg800_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg800_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1456 = reinterpret_tensor(buf1451, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1451 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_479], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1456, arg803_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg803_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1457 = reinterpret_tensor(buf1454, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1454 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_479], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1457, arg806_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg806_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_479], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1458 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1455, buf1456, buf1457, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1455 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1456 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1457 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1459 = buf1458[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1458 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1463 = buf1452; del buf1452 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_482], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1459, buf1463, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1459 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_482], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1464 = torch.ops.aten._weight_int4pack_mm.default(buf1463, arg807_1, 128, arg808_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg807_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg808_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1465 = buf1464 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1464 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1466 = reinterpret_tensor(buf1465, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1465 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_25, hidden_states_484, hidden_states_485], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1466, arg797_1, buf9, arg14_1, arg809_1, buf1441, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg809_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1441 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1467 = buf1463; del buf1463 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_102], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1466, buf1467, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_102], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1468 = torch.ops.aten._weight_int4pack_mm.default(buf1467, arg810_1, 128, arg811_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg810_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg811_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1469 = buf1468 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1468 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_102], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1471 = torch.ops.aten._weight_int4pack_mm.default(buf1470, arg813_1, 128, arg814_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg813_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg814_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1472 = buf1471 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1471 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_102], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1474 = torch.ops.aten._weight_int4pack_mm.default(buf1473, arg816_1, 128, arg817_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg816_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg817_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1475 = buf1474 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1474 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1476 = reinterpret_tensor(buf1469, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1469 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_486], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1476, arg812_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg812_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1477 = reinterpret_tensor(buf1472, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1472 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_486], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1477, arg815_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg815_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1478 = reinterpret_tensor(buf1475, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1475 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_486], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1478, arg818_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg818_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_486], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1480 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1476, buf1477, buf1478, reinterpret_tensor(buf1479, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1476 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1477 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1478 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1481 = buf1480[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1480 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1485 = buf1467; del buf1467 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_489], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1481, buf1485, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_489], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1486 = torch.ops.aten._weight_int4pack_mm.default(buf1485, arg819_1, 128, arg820_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg819_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg820_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1487 = buf1486 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1486 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1491 = reinterpret_tensor(buf1481, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1481 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_206, hidden_states_491, hidden_states_492, mul_106, norm_hidden_states_102, norm_hidden_states_103], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1487, arg821_1, buf1466, arg797_1, buf9, arg14_1, buf1491, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1492 = buf1485; del buf1485 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_493], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1491, buf1492, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_493], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1493 = torch.ops.aten._weight_int4pack_mm.default(buf1492, arg822_1, 128, arg823_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg822_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg823_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1494 = buf1493 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1493 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1495 = buf1438; del buf1438 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_496], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1494, arg824_1, buf1495, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg824_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1494 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_496], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1496 = torch.ops.aten._weight_int4pack_mm.default(buf1495, arg825_1, 128, arg826_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg825_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg826_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1497 = buf1496 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1496 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1498 = reinterpret_tensor(buf1497, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1497 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1502 = buf1491; del buf1491 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_210, ff_output_25, hidden_states_491, hidden_states_492, hidden_states_497, mul_108, norm_hidden_states_104, norm_hidden_states_105], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1498, arg797_1, buf9, arg14_1, arg827_1, buf1487, arg821_1, buf1466, arg828_1, buf1502, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg797_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg821_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg827_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1466 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1487 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1503 = buf1492; del buf1492 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1506 = buf1449; del buf1449 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1509 = buf1446; del buf1446 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_104, query_104, value_104], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1502, buf1503, buf1506, buf1509, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1502 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_104], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1504 = torch.ops.aten._weight_int4pack_mm.default(buf1503, arg829_1, 128, arg830_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg829_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg830_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1505 = buf1504 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1504 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_104], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1507 = torch.ops.aten._weight_int4pack_mm.default(buf1506, arg832_1, 128, arg833_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg832_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg833_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1508 = buf1507 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1507 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_104], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1510 = torch.ops.aten._weight_int4pack_mm.default(buf1509, arg835_1, 128, arg836_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg835_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg836_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1511 = buf1510 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1510 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1512 = reinterpret_tensor(buf1505, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1505 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_498], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1512, arg831_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg831_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1513 = reinterpret_tensor(buf1508, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1508 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_498], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1513, arg834_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg834_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1514 = reinterpret_tensor(buf1511, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1511 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_498], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1514, arg837_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg837_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_498], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1515 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1512, buf1513, buf1514, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1512 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1513 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1514 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1516 = buf1515[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1515 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1520 = buf1509; del buf1509 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_501], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1516, buf1520, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1516 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_501], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1521 = torch.ops.aten._weight_int4pack_mm.default(buf1520, arg838_1, 128, arg839_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg838_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg839_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1522 = buf1521 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1521 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1523 = reinterpret_tensor(buf1522, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1522 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_26, hidden_states_503, hidden_states_504], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1523, arg828_1, buf9, arg14_1, arg840_1, buf1498, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg840_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1498 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1524 = buf1520; del buf1520 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_106], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1523, buf1524, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_106], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1525 = torch.ops.aten._weight_int4pack_mm.default(buf1524, arg841_1, 128, arg842_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg841_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg842_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1526 = buf1525 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1525 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1527 = buf1473; del buf1473 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1530 = buf1470; del buf1470 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1584 = buf1416; del buf1416 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1587 = buf1413; del buf1413 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_106, key_110, value_106, value_110], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_12.run(buf44, arg21_1, buf1527, buf1530, buf1584, buf1587, 1228800, grid=grid(1228800), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg21_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf44 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_106], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1528 = torch.ops.aten._weight_int4pack_mm.default(buf1527, arg844_1, 128, arg845_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg844_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg845_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1527 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1529 = buf1528 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1528 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_106], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1531 = torch.ops.aten._weight_int4pack_mm.default(buf1530, arg847_1, 128, arg848_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg847_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg848_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1530 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1532 = buf1531 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1531 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1533 = reinterpret_tensor(buf1526, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1526 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_505], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1533, arg843_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg843_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1534 = reinterpret_tensor(buf1529, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1529 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_505], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1534, arg846_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg846_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1535 = reinterpret_tensor(buf1532, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1532 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_505], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1535, arg849_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg849_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1536 = buf1479; del buf1479 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1593 = buf1422; del buf1422 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_505, hidden_states_524], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_14.run(arg0_1, buf1536, buf1593, 9728, grid=grid(9728), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg0_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_505], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1537 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1533, buf1534, buf1535, reinterpret_tensor(buf1536, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1533 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1534 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1535 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1536 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1538 = buf1537[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1537 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1542 = buf1524; del buf1524 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_508], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1538, buf1542, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_508], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1543 = torch.ops.aten._weight_int4pack_mm.default(buf1542, arg850_1, 128, arg851_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg850_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg851_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1544 = buf1543 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1543 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1548 = reinterpret_tensor(buf1538, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1538 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_214, hidden_states_510, hidden_states_511, mul_110, norm_hidden_states_106, norm_hidden_states_107], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1544, arg852_1, buf1523, arg828_1, buf9, arg14_1, buf1548, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1549 = buf1542; del buf1542 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_512], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1548, buf1549, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_512], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1550 = torch.ops.aten._weight_int4pack_mm.default(buf1549, arg853_1, 128, arg854_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg853_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg854_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1551 = buf1550 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1550 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1552 = buf1495; del buf1495 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_515], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1551, arg855_1, buf1552, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg855_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1551 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_515], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1553 = torch.ops.aten._weight_int4pack_mm.default(buf1552, arg856_1, 128, arg857_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg856_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg857_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1554 = buf1553 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1553 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1555 = reinterpret_tensor(buf1554, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1554 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1559 = buf1548; del buf1548 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_218, ff_output_26, hidden_states_510, hidden_states_511, hidden_states_516, mul_112, norm_hidden_states_108, norm_hidden_states_109], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_17.run(buf1555, arg828_1, buf9, arg14_1, arg858_1, buf1544, arg852_1, buf1523, arg859_1, buf1559, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg828_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg852_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg858_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1523 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1544 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1560 = buf1549; del buf1549 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1563 = buf1506; del buf1506 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1566 = buf1503; del buf1503 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_108, query_108, value_108], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_6.run(buf1559, buf1560, buf1563, buf1566, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1559 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_108], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1561 = torch.ops.aten._weight_int4pack_mm.default(buf1560, arg860_1, 128, arg861_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg860_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg861_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1560 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1562 = buf1561 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1561 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_108], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1564 = torch.ops.aten._weight_int4pack_mm.default(buf1563, arg863_1, 128, arg864_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg863_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg864_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1563 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1565 = buf1564 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1564 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_108], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1567 = torch.ops.aten._weight_int4pack_mm.default(buf1566, arg866_1, 128, arg867_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg866_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg867_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1568 = buf1567 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1567 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1569 = reinterpret_tensor(buf1562, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1562 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_517], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1569, arg862_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg862_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1570 = reinterpret_tensor(buf1565, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1565 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_517], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1570, arg865_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg865_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1571 = reinterpret_tensor(buf1568, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1568 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_517], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1571, arg868_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg868_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_517], Original ATen: [aten._scaled_dot_product_cudnn_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1572 = torch.ops.aten._scaled_dot_product_cudnn_attention.default(buf1569, buf1570, buf1571, None, False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1569 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1570 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1571 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1573 = buf1572[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1572 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1577 = buf1566; del buf1566 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_520], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_8.run(buf1573, buf1577, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1573 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_520], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1578 = torch.ops.aten._weight_int4pack_mm.default(buf1577, arg869_1, 128, arg870_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg869_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg870_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1579 = buf1578 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1578 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1580 = reinterpret_tensor(buf1579, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1579 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [attn_output_27, hidden_states_522, hidden_states_523], Original ATen: [aten.add, aten.div, aten.mul] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_add_div_mul_18.run(buf1580, arg859_1, buf9, arg14_1, arg871_1, buf1555, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg871_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1555 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1581 = buf1577; del buf1577 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_110], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1580, buf1581, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [query_110], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1582 = torch.ops.aten._weight_int4pack_mm.default(buf1581, arg872_1, 128, arg873_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg872_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg873_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1583 = buf1582 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1582 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [key_110], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1585 = torch.ops.aten._weight_int4pack_mm.default(buf1584, arg875_1, 128, arg876_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg875_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg876_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1584 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1586 = buf1585 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1585 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [value_110], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1588 = torch.ops.aten._weight_int4pack_mm.default(buf1587, arg878_1, 128, arg879_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg878_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg879_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1587 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1589 = buf1588 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1588 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1590 = reinterpret_tensor(buf1583, (2, 16, 4096, 72), (4718592, 72, 1152, 1), 0); del buf1583 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_524], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_cudnn_attention_7.run(buf1590, arg874_1, 9437184, grid=grid(9437184), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg874_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1591 = reinterpret_tensor(buf1586, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1586 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_524], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1591, arg877_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg877_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1592 = reinterpret_tensor(buf1589, (2, 16, 300, 72), (345600, 72, 1152, 1), 0); del buf1589 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_524], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused__scaled_dot_product_efficient_attention_13.run(buf1592, arg880_1, 691200, grid=grid(691200), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg880_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_524], Original ATen: [aten._scaled_dot_product_efficient_attention] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1594 = torch.ops.aten._scaled_dot_product_efficient_attention.default(buf1590, buf1591, buf1592, reinterpret_tensor(buf1593, (2, 16, 4096, 300), (4864, 304, 0, 1), 0), False) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1590 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1591 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1592 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1593 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1595 = buf1594[0] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1594 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1599 = buf1581; del buf1581 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_527], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1595, buf1599, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_527], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1600 = torch.ops.aten._weight_int4pack_mm.default(buf1599, arg881_1, 128, arg882_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg881_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg882_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1601 = buf1600 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1600 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1605 = reinterpret_tensor(buf1595, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1595 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [add_222, hidden_states_529, hidden_states_530, mul_114, norm_hidden_states_110, norm_hidden_states_111], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_15.run(buf1601, arg883_1, buf1580, arg859_1, buf9, arg14_1, buf1605, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1606 = buf1599; del buf1599 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_531], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_10.run(buf1605, buf1606, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1605 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_531], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1607 = torch.ops.aten._weight_int4pack_mm.default(buf1606, arg884_1, 128, arg885_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg884_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg885_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1608 = buf1607 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1607 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1609 = buf1552; del buf1552 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_534], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_16.run(buf1608, arg886_1, buf1609, 41943040, grid=grid(41943040), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg886_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1608 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_534], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1610 = torch.ops.aten._weight_int4pack_mm.default(buf1609, arg887_1, 128, arg888_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg887_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg888_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1609 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1611 = buf1610 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1610 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1612 = reinterpret_tensor(buf1611, (2, 4096, 1152), (4718592, 1152, 1), 0); del buf1611 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1613 = empty_strided_cuda((2, 4096, 1), (4096, 1, 8192), torch.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1614 = empty_strided_cuda((2, 4096, 1), (4096, 1, 8192), torch.float32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [ff_output_27, hidden_states_529, hidden_states_530, hidden_states_535, hidden_states_536], Original ATen: [aten.add, aten.div, aten.mul, aten.native_layer_norm] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_red_fused_add_div_mul_native_layer_norm_19.run(buf1612, arg859_1, buf9, arg14_1, arg889_1, buf1601, arg883_1, buf1580, buf1613, buf1614, 8192, 1152, grid=grid(8192), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg14_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg859_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg883_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg889_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1580 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1601 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf9 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1616 = buf1606; del buf1606 # reuse V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_538], Original ATen: [aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_constant_pad_nd_20.run(buf1612, buf1613, buf1614, arg890_1, buf6, arg11_1, buf1616, 16777216, grid=grid(16777216), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg11_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg890_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1612 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1613 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1614 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf6 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [hidden_states_538], Original ATen: [aten._weight_int4pack_mm, aten.constant_pad_nd] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1617 = torch.ops.aten._weight_int4pack_mm.default(buf1616, arg891_1, 128, arg892_1) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg891_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg892_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1616 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1618 = buf1617 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1617 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] buf1619 = empty_strided_cuda((2, 8, 64, 2, 64, 2), (131072, 16384, 256, 128, 2, 1), torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] # Source Nodes: [output], Original ATen: [aten.clone] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] triton_poi_fused_clone_21.run(buf1618, arg893_1, buf1619, 16, 16384, grid=grid(16, 16384), stream=stream0) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del arg893_1 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] del buf1618 V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] return (reinterpret_tensor(buf1619, (2, 8, 128, 128), (131072, 16384, 128, 1), 0), ) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] def benchmark_compiled_module(times=10, repeat=10): V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._dynamo.testing import rand_strided V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.utils import print_performance V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg0_1 = rand_strided((2, 300), (300, 1), device='cuda:0', dtype=torch.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg1_1 = rand_strided((2, 4, 128, 128), (65536, 16384, 128, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg2_1 = rand_strided((1152, 4, 2, 2), (16, 1, 8, 4), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg3_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg4_1 = rand_strided((1, 4096, 1152), (4718592, 1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg5_1 = rand_strided((2, ), (0, ), device='cuda:0', dtype=torch.int64) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg6_1 = rand_strided((144, 8, 32, 4), (1024, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg7_1 = rand_strided((8, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg8_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg9_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg10_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg11_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg12_1 = rand_strided((864, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg13_1 = rand_strided((16, 6912, 2), (13824, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg14_1 = rand_strided((6912, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg15_1 = rand_strided((144, 32, 32, 4), (4096, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg16_1 = rand_strided((32, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg17_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg18_1 = rand_strided((2, 300, 4096), (1228800, 4096, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg19_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg20_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg21_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg22_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg23_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg24_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg25_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg26_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg27_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg28_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg29_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg30_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg31_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg32_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg33_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg34_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg35_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg36_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg37_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg38_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg39_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg40_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg41_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg42_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg43_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg44_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg45_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg46_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg47_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg48_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg49_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg50_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg51_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg52_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg53_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg54_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg55_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg56_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg57_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg58_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg59_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg60_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg61_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg62_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg63_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg64_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg65_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg66_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg67_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg68_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg69_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg70_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg71_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg72_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg73_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg74_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg75_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg76_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg77_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg78_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg79_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg80_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg81_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg82_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg83_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg84_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg85_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg86_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg87_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg88_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg89_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg90_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg91_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg92_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg93_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg94_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg95_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg96_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg97_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg98_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg99_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg100_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg101_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg102_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg103_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg104_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg105_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg106_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg107_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg108_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg109_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg110_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg111_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg112_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg113_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg114_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg115_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg116_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg117_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg118_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg119_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg120_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg121_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg122_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg123_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg124_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg125_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg126_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg127_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg128_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg129_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg130_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg131_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg132_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg133_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg134_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg135_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg136_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg137_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg138_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg139_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg140_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg141_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg142_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg143_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg144_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg145_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg146_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg147_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg148_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg149_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg150_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg151_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg152_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg153_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg154_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg155_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg156_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg157_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg158_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg159_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg160_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg161_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg162_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg163_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg164_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg165_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg166_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg167_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg168_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg169_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg170_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg171_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg172_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg173_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg174_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg175_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg176_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg177_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg178_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg179_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg180_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg181_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg182_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg183_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg184_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg185_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg186_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg187_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg188_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg189_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg190_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg191_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg192_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg193_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg194_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg195_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg196_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg197_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg198_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg199_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg200_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg201_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg202_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg203_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg204_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg205_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg206_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg207_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg208_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg209_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg210_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg211_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg212_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg213_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg214_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg215_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg216_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg217_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg218_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg219_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg220_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg221_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg222_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg223_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg224_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg225_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg226_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg227_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg228_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg229_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg230_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg231_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg232_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg233_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg234_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg235_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg236_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg237_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg238_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg239_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg240_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg241_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg242_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg243_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg244_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg245_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg246_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg247_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg248_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg249_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg250_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg251_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg252_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg253_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg254_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg255_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg256_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg257_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg258_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg259_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg260_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg261_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg262_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg263_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg264_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg265_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg266_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg267_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg268_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg269_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg270_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg271_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg272_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg273_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg274_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg275_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg276_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg277_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg278_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg279_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg280_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg281_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg282_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg283_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg284_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg285_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg286_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg287_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg288_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg289_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg290_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg291_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg292_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg293_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg294_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg295_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg296_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg297_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg298_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg299_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg300_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg301_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg302_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg303_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg304_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg305_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg306_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg307_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg308_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg309_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg310_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg311_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg312_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg313_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg314_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg315_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg316_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg317_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg318_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg319_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg320_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg321_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg322_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg323_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg324_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg325_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg326_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg327_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg328_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg329_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg330_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg331_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg332_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg333_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg334_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg335_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg336_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg337_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg338_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg339_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg340_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg341_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg342_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg343_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg344_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg345_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg346_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg347_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg348_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg349_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg350_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg351_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg352_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg353_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg354_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg355_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg356_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg357_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg358_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg359_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg360_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg361_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg362_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg363_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg364_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg365_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg366_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg367_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg368_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg369_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg370_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg371_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg372_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg373_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg374_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg375_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg376_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg377_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg378_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg379_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg380_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg381_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg382_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg383_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg384_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg385_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg386_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg387_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg388_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg389_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg390_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg391_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg392_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg393_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg394_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg395_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg396_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg397_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg398_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg399_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg400_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg401_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg402_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg403_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg404_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg405_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg406_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg407_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg408_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg409_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg410_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg411_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg412_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg413_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg414_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg415_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg416_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg417_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg418_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg419_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg420_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg421_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg422_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg423_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg424_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg425_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg426_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg427_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg428_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg429_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg430_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg431_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg432_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg433_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg434_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg435_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg436_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg437_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg438_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg439_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg440_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg441_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg442_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg443_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg444_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg445_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg446_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg447_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg448_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg449_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg450_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg451_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg452_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg453_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg454_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg455_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg456_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg457_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg458_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg459_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg460_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg461_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg462_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg463_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg464_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg465_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg466_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg467_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg468_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg469_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg470_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg471_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg472_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg473_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg474_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg475_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg476_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg477_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg478_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg479_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg480_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg481_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg482_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg483_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg484_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg485_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg486_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg487_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg488_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg489_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg490_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg491_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg492_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg493_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg494_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg495_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg496_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg497_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg498_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg499_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg500_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg501_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg502_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg503_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg504_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg505_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg506_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg507_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg508_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg509_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg510_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg511_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg512_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg513_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg514_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg515_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg516_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg517_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg518_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg519_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg520_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg521_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg522_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg523_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg524_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg525_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg526_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg527_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg528_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg529_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg530_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg531_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg532_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg533_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg534_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg535_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg536_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg537_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg538_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg539_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg540_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg541_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg542_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg543_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg544_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg545_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg546_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg547_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg548_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg549_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg550_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg551_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg552_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg553_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg554_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg555_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg556_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg557_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg558_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg559_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg560_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg561_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg562_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg563_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg564_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg565_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg566_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg567_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg568_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg569_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg570_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg571_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg572_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg573_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg574_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg575_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg576_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg577_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg578_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg579_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg580_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg581_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg582_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg583_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg584_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg585_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg586_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg587_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg588_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg589_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg590_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg591_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg592_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg593_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg594_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg595_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg596_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg597_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg598_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg599_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg600_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg601_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg602_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg603_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg604_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg605_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg606_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg607_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg608_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg609_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg610_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg611_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg612_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg613_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg614_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg615_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg616_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg617_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg618_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg619_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg620_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg621_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg622_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg623_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg624_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg625_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg626_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg627_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg628_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg629_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg630_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg631_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg632_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg633_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg634_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg635_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg636_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg637_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg638_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg639_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg640_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg641_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg642_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg643_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg644_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg645_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg646_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg647_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg648_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg649_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg650_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg651_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg652_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg653_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg654_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg655_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg656_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg657_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg658_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg659_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg660_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg661_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg662_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg663_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg664_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg665_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg666_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg667_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg668_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg669_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg670_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg671_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg672_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg673_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg674_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg675_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg676_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg677_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg678_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg679_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg680_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg681_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg682_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg683_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg684_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg685_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg686_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg687_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg688_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg689_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg690_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg691_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg692_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg693_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg694_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg695_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg696_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg697_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg698_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg699_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg700_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg701_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg702_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg703_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg704_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg705_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg706_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg707_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg708_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg709_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg710_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg711_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg712_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg713_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg714_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg715_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg716_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg717_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg718_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg719_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg720_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg721_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg722_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg723_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg724_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg725_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg726_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg727_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg728_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg729_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg730_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg731_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg732_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg733_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg734_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg735_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg736_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg737_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg738_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg739_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg740_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg741_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg742_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg743_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg744_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg745_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg746_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg747_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg748_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg749_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg750_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg751_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg752_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg753_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg754_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg755_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg756_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg757_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg758_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg759_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg760_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg761_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg762_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg763_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg764_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg765_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg766_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg767_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg768_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg769_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg770_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg771_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg772_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg773_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg774_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg775_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg776_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg777_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg778_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg779_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg780_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg781_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg782_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg783_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg784_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg785_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg786_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg787_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg788_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg789_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg790_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg791_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg792_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg793_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg794_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg795_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg796_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg797_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg798_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg799_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg800_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg801_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg802_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg803_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg804_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg805_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg806_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg807_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg808_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg809_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg810_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg811_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg812_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg813_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg814_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg815_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg816_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg817_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg818_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg819_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg820_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg821_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg822_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg823_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg824_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg825_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg826_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg827_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg828_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg829_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg830_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg831_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg832_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg833_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg834_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg835_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg836_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg837_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg838_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg839_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg840_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg841_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg842_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg843_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg844_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg845_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg846_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg847_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg848_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg849_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg850_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg851_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg852_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg853_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg854_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg855_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg856_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg857_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg858_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg859_1 = rand_strided((6, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg860_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg861_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg862_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg863_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg864_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg865_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg866_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg867_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg868_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg869_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg870_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg871_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg872_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg873_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg874_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg875_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg876_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg877_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg878_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg879_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg880_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg881_1 = rand_strided((144, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg882_1 = rand_strided((16, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg883_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg884_1 = rand_strided((576, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg885_1 = rand_strided((16, 4608, 2), (9216, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg886_1 = rand_strided((4608, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg887_1 = rand_strided((144, 40, 32, 4), (5120, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg888_1 = rand_strided((40, 1152, 2), (2304, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg889_1 = rand_strided((1152, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg890_1 = rand_strided((2, 1152), (1152, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg891_1 = rand_strided((4, 16, 32, 4), (2048, 128, 4, 1), device='cuda:0', dtype=torch.int32) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg892_1 = rand_strided((16, 32, 2), (64, 2, 1), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] arg893_1 = rand_strided((32, ), (1, ), device='cuda:0', dtype=torch.bfloat16) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] fn = lambda: call([arg0_1, arg1_1, arg2_1, arg3_1, arg4_1, arg5_1, arg6_1, arg7_1, arg8_1, arg9_1, arg10_1, arg11_1, arg12_1, arg13_1, arg14_1, arg15_1, arg16_1, arg17_1, arg18_1, arg19_1, arg20_1, arg21_1, arg22_1, arg23_1, arg24_1, arg25_1, arg26_1, arg27_1, arg28_1, arg29_1, arg30_1, arg31_1, arg32_1, arg33_1, arg34_1, arg35_1, arg36_1, arg37_1, arg38_1, arg39_1, arg40_1, arg41_1, arg42_1, arg43_1, arg44_1, arg45_1, arg46_1, arg47_1, arg48_1, arg49_1, arg50_1, arg51_1, arg52_1, arg53_1, arg54_1, arg55_1, arg56_1, arg57_1, arg58_1, arg59_1, arg60_1, arg61_1, arg62_1, arg63_1, arg64_1, arg65_1, arg66_1, arg67_1, arg68_1, arg69_1, arg70_1, arg71_1, arg72_1, arg73_1, arg74_1, arg75_1, arg76_1, arg77_1, arg78_1, arg79_1, arg80_1, arg81_1, arg82_1, arg83_1, arg84_1, arg85_1, arg86_1, arg87_1, arg88_1, arg89_1, arg90_1, arg91_1, arg92_1, arg93_1, arg94_1, arg95_1, arg96_1, arg97_1, arg98_1, arg99_1, arg100_1, arg101_1, arg102_1, arg103_1, arg104_1, arg105_1, arg106_1, arg107_1, arg108_1, arg109_1, arg110_1, arg111_1, arg112_1, arg113_1, arg114_1, arg115_1, arg116_1, arg117_1, arg118_1, arg119_1, arg120_1, arg121_1, arg122_1, arg123_1, arg124_1, arg125_1, arg126_1, arg127_1, arg128_1, arg129_1, arg130_1, arg131_1, arg132_1, arg133_1, arg134_1, arg135_1, arg136_1, arg137_1, arg138_1, arg139_1, arg140_1, arg141_1, arg142_1, arg143_1, arg144_1, arg145_1, arg146_1, arg147_1, arg148_1, arg149_1, arg150_1, arg151_1, arg152_1, arg153_1, arg154_1, arg155_1, arg156_1, arg157_1, arg158_1, arg159_1, arg160_1, arg161_1, arg162_1, arg163_1, arg164_1, arg165_1, arg166_1, arg167_1, arg168_1, arg169_1, arg170_1, arg171_1, arg172_1, arg173_1, arg174_1, arg175_1, arg176_1, arg177_1, arg178_1, arg179_1, arg180_1, arg181_1, arg182_1, arg183_1, arg184_1, arg185_1, arg186_1, arg187_1, arg188_1, arg189_1, arg190_1, arg191_1, arg192_1, arg193_1, arg194_1, arg195_1, arg196_1, arg197_1, arg198_1, arg199_1, arg200_1, arg201_1, arg202_1, arg203_1, arg204_1, arg205_1, arg206_1, arg207_1, arg208_1, arg209_1, arg210_1, arg211_1, arg212_1, arg213_1, arg214_1, arg215_1, arg216_1, arg217_1, arg218_1, arg219_1, arg220_1, arg221_1, arg222_1, arg223_1, arg224_1, arg225_1, arg226_1, arg227_1, arg228_1, arg229_1, arg230_1, arg231_1, arg232_1, arg233_1, arg234_1, arg235_1, arg236_1, arg237_1, arg238_1, arg239_1, arg240_1, arg241_1, arg242_1, arg243_1, arg244_1, arg245_1, arg246_1, arg247_1, arg248_1, arg249_1, arg250_1, arg251_1, arg252_1, arg253_1, arg254_1, arg255_1, arg256_1, arg257_1, arg258_1, arg259_1, arg260_1, arg261_1, arg262_1, arg263_1, arg264_1, arg265_1, arg266_1, arg267_1, arg268_1, arg269_1, arg270_1, arg271_1, arg272_1, arg273_1, arg274_1, arg275_1, arg276_1, arg277_1, arg278_1, arg279_1, arg280_1, arg281_1, arg282_1, arg283_1, arg284_1, arg285_1, arg286_1, arg287_1, arg288_1, arg289_1, arg290_1, arg291_1, arg292_1, arg293_1, arg294_1, arg295_1, arg296_1, arg297_1, arg298_1, arg299_1, arg300_1, arg301_1, arg302_1, arg303_1, arg304_1, arg305_1, arg306_1, arg307_1, arg308_1, arg309_1, arg310_1, arg311_1, arg312_1, arg313_1, arg314_1, arg315_1, arg316_1, arg317_1, arg318_1, arg319_1, arg320_1, arg321_1, arg322_1, arg323_1, arg324_1, arg325_1, arg326_1, arg327_1, arg328_1, arg329_1, arg330_1, arg331_1, arg332_1, arg333_1, arg334_1, arg335_1, arg336_1, arg337_1, arg338_1, arg339_1, arg340_1, arg341_1, arg342_1, arg343_1, arg344_1, arg345_1, arg346_1, arg347_1, arg348_1, arg349_1, arg350_1, arg351_1, arg352_1, arg353_1, arg354_1, arg355_1, arg356_1, arg357_1, arg358_1, arg359_1, arg360_1, arg361_1, arg362_1, arg363_1, arg364_1, arg365_1, arg366_1, arg367_1, arg368_1, arg369_1, arg370_1, arg371_1, arg372_1, arg373_1, arg374_1, arg375_1, arg376_1, arg377_1, arg378_1, arg379_1, arg380_1, arg381_1, arg382_1, arg383_1, arg384_1, arg385_1, arg386_1, arg387_1, arg388_1, arg389_1, arg390_1, arg391_1, arg392_1, arg393_1, arg394_1, arg395_1, arg396_1, arg397_1, arg398_1, arg399_1, arg400_1, arg401_1, arg402_1, arg403_1, arg404_1, arg405_1, arg406_1, arg407_1, arg408_1, arg409_1, arg410_1, arg411_1, arg412_1, arg413_1, arg414_1, arg415_1, arg416_1, arg417_1, arg418_1, arg419_1, arg420_1, arg421_1, arg422_1, arg423_1, arg424_1, arg425_1, arg426_1, arg427_1, arg428_1, arg429_1, arg430_1, arg431_1, arg432_1, arg433_1, arg434_1, arg435_1, arg436_1, arg437_1, arg438_1, arg439_1, arg440_1, arg441_1, arg442_1, arg443_1, arg444_1, arg445_1, arg446_1, arg447_1, arg448_1, arg449_1, arg450_1, arg451_1, arg452_1, arg453_1, arg454_1, arg455_1, arg456_1, arg457_1, arg458_1, arg459_1, arg460_1, arg461_1, arg462_1, arg463_1, arg464_1, arg465_1, arg466_1, arg467_1, arg468_1, arg469_1, arg470_1, arg471_1, arg472_1, arg473_1, arg474_1, arg475_1, arg476_1, arg477_1, arg478_1, arg479_1, arg480_1, arg481_1, arg482_1, arg483_1, arg484_1, arg485_1, arg486_1, arg487_1, arg488_1, arg489_1, arg490_1, arg491_1, arg492_1, arg493_1, arg494_1, arg495_1, arg496_1, arg497_1, arg498_1, arg499_1, arg500_1, arg501_1, arg502_1, arg503_1, arg504_1, arg505_1, arg506_1, arg507_1, arg508_1, arg509_1, arg510_1, arg511_1, arg512_1, arg513_1, arg514_1, arg515_1, arg516_1, arg517_1, arg518_1, arg519_1, arg520_1, arg521_1, arg522_1, arg523_1, arg524_1, arg525_1, arg526_1, arg527_1, arg528_1, arg529_1, arg530_1, arg531_1, arg532_1, arg533_1, arg534_1, arg535_1, arg536_1, arg537_1, arg538_1, arg539_1, arg540_1, arg541_1, arg542_1, arg543_1, arg544_1, arg545_1, arg546_1, arg547_1, arg548_1, arg549_1, arg550_1, arg551_1, arg552_1, arg553_1, arg554_1, arg555_1, arg556_1, arg557_1, arg558_1, arg559_1, arg560_1, arg561_1, arg562_1, arg563_1, arg564_1, arg565_1, arg566_1, arg567_1, arg568_1, arg569_1, arg570_1, arg571_1, arg572_1, arg573_1, arg574_1, arg575_1, arg576_1, arg577_1, arg578_1, arg579_1, arg580_1, arg581_1, arg582_1, arg583_1, arg584_1, arg585_1, arg586_1, arg587_1, arg588_1, arg589_1, arg590_1, arg591_1, arg592_1, arg593_1, arg594_1, arg595_1, arg596_1, arg597_1, arg598_1, arg599_1, arg600_1, arg601_1, arg602_1, arg603_1, arg604_1, arg605_1, arg606_1, arg607_1, arg608_1, arg609_1, arg610_1, arg611_1, arg612_1, arg613_1, arg614_1, arg615_1, arg616_1, arg617_1, arg618_1, arg619_1, arg620_1, arg621_1, arg622_1, arg623_1, arg624_1, arg625_1, arg626_1, arg627_1, arg628_1, arg629_1, arg630_1, arg631_1, arg632_1, arg633_1, arg634_1, arg635_1, arg636_1, arg637_1, arg638_1, arg639_1, arg640_1, arg641_1, arg642_1, arg643_1, arg644_1, arg645_1, arg646_1, arg647_1, arg648_1, arg649_1, arg650_1, arg651_1, arg652_1, arg653_1, arg654_1, arg655_1, arg656_1, arg657_1, arg658_1, arg659_1, arg660_1, arg661_1, arg662_1, arg663_1, arg664_1, arg665_1, arg666_1, arg667_1, arg668_1, arg669_1, arg670_1, arg671_1, arg672_1, arg673_1, arg674_1, arg675_1, arg676_1, arg677_1, arg678_1, arg679_1, arg680_1, arg681_1, arg682_1, arg683_1, arg684_1, arg685_1, arg686_1, arg687_1, arg688_1, arg689_1, arg690_1, arg691_1, arg692_1, arg693_1, arg694_1, arg695_1, arg696_1, arg697_1, arg698_1, arg699_1, arg700_1, arg701_1, arg702_1, arg703_1, arg704_1, arg705_1, arg706_1, arg707_1, arg708_1, arg709_1, arg710_1, arg711_1, arg712_1, arg713_1, arg714_1, arg715_1, arg716_1, arg717_1, arg718_1, arg719_1, arg720_1, arg721_1, arg722_1, arg723_1, arg724_1, arg725_1, arg726_1, arg727_1, arg728_1, arg729_1, arg730_1, arg731_1, arg732_1, arg733_1, arg734_1, arg735_1, arg736_1, arg737_1, arg738_1, arg739_1, arg740_1, arg741_1, arg742_1, arg743_1, arg744_1, arg745_1, arg746_1, arg747_1, arg748_1, arg749_1, arg750_1, arg751_1, arg752_1, arg753_1, arg754_1, arg755_1, arg756_1, arg757_1, arg758_1, arg759_1, arg760_1, arg761_1, arg762_1, arg763_1, arg764_1, arg765_1, arg766_1, arg767_1, arg768_1, arg769_1, arg770_1, arg771_1, arg772_1, arg773_1, arg774_1, arg775_1, arg776_1, arg777_1, arg778_1, arg779_1, arg780_1, arg781_1, arg782_1, arg783_1, arg784_1, arg785_1, arg786_1, arg787_1, arg788_1, arg789_1, arg790_1, arg791_1, arg792_1, arg793_1, arg794_1, arg795_1, arg796_1, arg797_1, arg798_1, arg799_1, arg800_1, arg801_1, arg802_1, arg803_1, arg804_1, arg805_1, arg806_1, arg807_1, arg808_1, arg809_1, arg810_1, arg811_1, arg812_1, arg813_1, arg814_1, arg815_1, arg816_1, arg817_1, arg818_1, arg819_1, arg820_1, arg821_1, arg822_1, arg823_1, arg824_1, arg825_1, arg826_1, arg827_1, arg828_1, arg829_1, arg830_1, arg831_1, arg832_1, arg833_1, arg834_1, arg835_1, arg836_1, arg837_1, arg838_1, arg839_1, arg840_1, arg841_1, arg842_1, arg843_1, arg844_1, arg845_1, arg846_1, arg847_1, arg848_1, arg849_1, arg850_1, arg851_1, arg852_1, arg853_1, arg854_1, arg855_1, arg856_1, arg857_1, arg858_1, arg859_1, arg860_1, arg861_1, arg862_1, arg863_1, arg864_1, arg865_1, arg866_1, arg867_1, arg868_1, arg869_1, arg870_1, arg871_1, arg872_1, arg873_1, arg874_1, arg875_1, arg876_1, arg877_1, arg878_1, arg879_1, arg880_1, arg881_1, arg882_1, arg883_1, arg884_1, arg885_1, arg886_1, arg887_1, arg888_1, arg889_1, arg890_1, arg891_1, arg892_1, arg893_1]) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] return print_performance(fn, times=times, repeat=repeat) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] if __name__ == "__main__": V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] from torch._inductor.wrapper_benchmark import compiled_module_main V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] compiled_module_main('None', benchmark_compiled_module) V0808 07:30:27.814042 1266882 torch/_inductor/graph.py:1780] [0/0] [__output_code] I0808 07:30:31.449268 1266882 torch/_inductor/graph.py:1814] [0/0] [__output_code] Output code written to: /tmp/torchinductor_sayak/zb/czbvhwljs3kusciezfqmyq4h44hfd5dma5sfliqylshsi5zrrsli.py