diff --git "a/nemotron_coreml_160ms/encoder/encoder_int8.mlmodelc/model.mil" "b/nemotron_coreml_160ms/encoder/encoder_int8.mlmodelc/model.mil" new file mode 100644--- /dev/null +++ "b/nemotron_coreml_160ms/encoder/encoder_int8.mlmodelc/model.mil" @@ -0,0 +1,4190 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}})] +{ + func main(tensor cache_channel, tensor cache_len, tensor cache_time, tensor mel, tensor mel_length) { + tensor module_pre_encode_conv_0_bias = const()[name = tensor("module_pre_encode_conv_0_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor module_pre_encode_conv_0_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_0_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3840))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3520)))]; + tensor module_pre_encode_conv_2_bias = const()[name = tensor("module_pre_encode_conv_2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4928)))]; + tensor module_pre_encode_conv_2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(6016))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8704))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8384)))]; + tensor module_pre_encode_conv_3_bias = const()[name = tensor("module_pre_encode_conv_3_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9792)))]; + tensor module_pre_encode_conv_3_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_3_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10880))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76800))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76480)))]; + tensor module_pre_encode_conv_5_bias = const()[name = tensor("module_pre_encode_conv_5_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77888)))]; + tensor module_pre_encode_conv_5_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_5_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78976))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81344)))]; + tensor module_pre_encode_conv_6_bias = const()[name = tensor("module_pre_encode_conv_6_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82752)))]; + tensor module_pre_encode_conv_6_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_conv_6_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149760))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149440)))]; + tensor module_pre_encode_out_bias = const()[name = tensor("module_pre_encode_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150848)))]; + tensor module_pre_encode_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_pre_encode_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155008))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4612608))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4611520)))]; + tensor module_layers_0_norm_feed_forward1_bias = const()[name = tensor("module_layers_0_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4616768)))]; + tensor module_layers_0_norm_feed_forward1_weight = const()[name = tensor("module_layers_0_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4620928)))]; + tensor module_layers_0_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4625088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8823616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8819456)))]; + tensor module_layers_0_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8840064))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13035520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13034432)))]; + tensor module_layers_0_norm_self_att_bias = const()[name = tensor("module_layers_0_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13039680)))]; + tensor module_layers_0_norm_self_att_weight = const()[name = tensor("module_layers_0_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13043840)))]; + tensor module_layers_0_self_attn_pos_bias_v = const()[name = tensor("module_layers_0_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13048000)))]; + tensor module_layers_0_self_attn_pos_bias_u = const()[name = tensor("module_layers_0_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13052160)))]; + tensor module_layers_0_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13056320))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14106048))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14104960)))]; + tensor module_layers_0_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14110208))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15159936))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15158848)))]; + tensor module_layers_0_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15164096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16213824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16212736)))]; + tensor module_layers_0_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16217984))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17267712))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17266624)))]; + tensor module_layers_0_norm_conv_bias = const()[name = tensor("module_layers_0_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17271872)))]; + tensor module_layers_0_norm_conv_weight = const()[name = tensor("module_layers_0_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17276032)))]; + tensor module_layers_0_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17280192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19379520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19377408)))]; + tensor module_layers_0_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19387776))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19398144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19397056)))]; + tensor module_layers_0_conv_batch_norm_bias = const()[name = tensor("module_layers_0_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19402304)))]; + tensor module_layers_0_conv_batch_norm_weight = const()[name = tensor("module_layers_0_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19406464)))]; + tensor module_layers_0_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19410624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20460352))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20459264)))]; + tensor module_layers_0_norm_feed_forward2_bias = const()[name = tensor("module_layers_0_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20464512)))]; + tensor module_layers_0_norm_feed_forward2_weight = const()[name = tensor("module_layers_0_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20468672)))]; + tensor module_layers_0_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20472832))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24671360))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24667200)))]; + tensor module_layers_0_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_0_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24687808))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28883264))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28882176)))]; + tensor module_layers_0_norm_out_bias = const()[name = tensor("module_layers_0_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28887424)))]; + tensor module_layers_0_norm_out_weight = const()[name = tensor("module_layers_0_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28891584)))]; + tensor module_layers_1_norm_feed_forward1_bias = const()[name = tensor("module_layers_1_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28895744)))]; + tensor module_layers_1_norm_feed_forward1_weight = const()[name = tensor("module_layers_1_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28899904)))]; + tensor module_layers_1_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28904064))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33102592))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33098432)))]; + tensor module_layers_1_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33119040))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37314496))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37313408)))]; + tensor module_layers_1_norm_self_att_bias = const()[name = tensor("module_layers_1_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37318656)))]; + tensor module_layers_1_norm_self_att_weight = const()[name = tensor("module_layers_1_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37322816)))]; + tensor module_layers_1_self_attn_pos_bias_v = const()[name = tensor("module_layers_1_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37326976)))]; + tensor module_layers_1_self_attn_pos_bias_u = const()[name = tensor("module_layers_1_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37331136)))]; + tensor module_layers_1_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(37335296))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38385024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38383936)))]; + tensor module_layers_1_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38389184))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39438912))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39437824)))]; + tensor module_layers_1_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39443072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40492800))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40491712)))]; + tensor module_layers_1_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40496960))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41546688))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41545600)))]; + tensor module_layers_1_norm_conv_bias = const()[name = tensor("module_layers_1_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41550848)))]; + tensor module_layers_1_norm_conv_weight = const()[name = tensor("module_layers_1_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41555008)))]; + tensor module_layers_1_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41559168))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43658496))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43656384)))]; + tensor module_layers_1_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43666752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43677120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43676032)))]; + tensor module_layers_1_conv_batch_norm_bias = const()[name = tensor("module_layers_1_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43681280)))]; + tensor module_layers_1_conv_batch_norm_weight = const()[name = tensor("module_layers_1_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43685440)))]; + tensor module_layers_1_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43689600))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44739328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44738240)))]; + tensor module_layers_1_norm_feed_forward2_bias = const()[name = tensor("module_layers_1_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44743488)))]; + tensor module_layers_1_norm_feed_forward2_weight = const()[name = tensor("module_layers_1_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44747648)))]; + tensor module_layers_1_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44751808))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48950336))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48946176)))]; + tensor module_layers_1_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_1_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48966784))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53162240))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53161152)))]; + tensor module_layers_1_norm_out_bias = const()[name = tensor("module_layers_1_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53166400)))]; + tensor module_layers_1_norm_out_weight = const()[name = tensor("module_layers_1_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53170560)))]; + tensor module_layers_2_norm_feed_forward1_bias = const()[name = tensor("module_layers_2_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53174720)))]; + tensor module_layers_2_norm_feed_forward1_weight = const()[name = tensor("module_layers_2_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53178880)))]; + tensor module_layers_2_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53183040))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57381568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57377408)))]; + tensor module_layers_2_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57398016))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61593472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61592384)))]; + tensor module_layers_2_norm_self_att_bias = const()[name = tensor("module_layers_2_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61597632)))]; + tensor module_layers_2_norm_self_att_weight = const()[name = tensor("module_layers_2_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61601792)))]; + tensor module_layers_2_self_attn_pos_bias_v = const()[name = tensor("module_layers_2_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61605952)))]; + tensor module_layers_2_self_attn_pos_bias_u = const()[name = tensor("module_layers_2_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61610112)))]; + tensor module_layers_2_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61614272))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62664000))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62662912)))]; + tensor module_layers_2_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(62668160))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63717888))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63716800)))]; + tensor module_layers_2_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63722048))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64771776))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64770688)))]; + tensor module_layers_2_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64775936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65825664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65824576)))]; + tensor module_layers_2_norm_conv_bias = const()[name = tensor("module_layers_2_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65829824)))]; + tensor module_layers_2_norm_conv_weight = const()[name = tensor("module_layers_2_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65833984)))]; + tensor module_layers_2_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65838144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67937472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67935360)))]; + tensor module_layers_2_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67945728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67956096))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67955008)))]; + tensor module_layers_2_conv_batch_norm_bias = const()[name = tensor("module_layers_2_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67960256)))]; + tensor module_layers_2_conv_batch_norm_weight = const()[name = tensor("module_layers_2_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67964416)))]; + tensor module_layers_2_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67968576))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69018304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69017216)))]; + tensor module_layers_2_norm_feed_forward2_bias = const()[name = tensor("module_layers_2_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69022464)))]; + tensor module_layers_2_norm_feed_forward2_weight = const()[name = tensor("module_layers_2_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69026624)))]; + tensor module_layers_2_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69030784))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73229312))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73225152)))]; + tensor module_layers_2_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_2_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73245760))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77441216))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77440128)))]; + tensor module_layers_2_norm_out_bias = const()[name = tensor("module_layers_2_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77445376)))]; + tensor module_layers_2_norm_out_weight = const()[name = tensor("module_layers_2_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77449536)))]; + tensor module_layers_3_norm_feed_forward1_bias = const()[name = tensor("module_layers_3_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77453696)))]; + tensor module_layers_3_norm_feed_forward1_weight = const()[name = tensor("module_layers_3_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77457856)))]; + tensor module_layers_3_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77462016))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81660544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81656384)))]; + tensor module_layers_3_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(81676992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85872448))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85871360)))]; + tensor module_layers_3_norm_self_att_bias = const()[name = tensor("module_layers_3_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85876608)))]; + tensor module_layers_3_norm_self_att_weight = const()[name = tensor("module_layers_3_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85880768)))]; + tensor module_layers_3_self_attn_pos_bias_v = const()[name = tensor("module_layers_3_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85884928)))]; + tensor module_layers_3_self_attn_pos_bias_u = const()[name = tensor("module_layers_3_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85889088)))]; + tensor module_layers_3_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85893248))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86942976))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86941888)))]; + tensor module_layers_3_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86947136))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87996864))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(87995776)))]; + tensor module_layers_3_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88001024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89050752))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89049664)))]; + tensor module_layers_3_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89054912))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90104640))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90103552)))]; + tensor module_layers_3_norm_conv_bias = const()[name = tensor("module_layers_3_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90108800)))]; + tensor module_layers_3_norm_conv_weight = const()[name = tensor("module_layers_3_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90112960)))]; + tensor module_layers_3_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90117120))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92216448))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92214336)))]; + tensor module_layers_3_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92224704))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92235072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92233984)))]; + tensor module_layers_3_conv_batch_norm_bias = const()[name = tensor("module_layers_3_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92239232)))]; + tensor module_layers_3_conv_batch_norm_weight = const()[name = tensor("module_layers_3_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92243392)))]; + tensor module_layers_3_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92247552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93297280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93296192)))]; + tensor module_layers_3_norm_feed_forward2_bias = const()[name = tensor("module_layers_3_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93301440)))]; + tensor module_layers_3_norm_feed_forward2_weight = const()[name = tensor("module_layers_3_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93305600)))]; + tensor module_layers_3_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93309760))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97508288))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97504128)))]; + tensor module_layers_3_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_3_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97524736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101720192))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101719104)))]; + tensor module_layers_3_norm_out_bias = const()[name = tensor("module_layers_3_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101724352)))]; + tensor module_layers_3_norm_out_weight = const()[name = tensor("module_layers_3_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101728512)))]; + tensor module_layers_4_norm_feed_forward1_bias = const()[name = tensor("module_layers_4_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101732672)))]; + tensor module_layers_4_norm_feed_forward1_weight = const()[name = tensor("module_layers_4_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101736832)))]; + tensor module_layers_4_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101740992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105939520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105935360)))]; + tensor module_layers_4_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105955968))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110151424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110150336)))]; + tensor module_layers_4_norm_self_att_bias = const()[name = tensor("module_layers_4_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110155584)))]; + tensor module_layers_4_norm_self_att_weight = const()[name = tensor("module_layers_4_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110159744)))]; + tensor module_layers_4_self_attn_pos_bias_v = const()[name = tensor("module_layers_4_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110163904)))]; + tensor module_layers_4_self_attn_pos_bias_u = const()[name = tensor("module_layers_4_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110168064)))]; + tensor module_layers_4_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110172224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111221952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111220864)))]; + tensor module_layers_4_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111226112))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112275840))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112274752)))]; + tensor module_layers_4_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(112280000))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113329728))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113328640)))]; + tensor module_layers_4_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113333888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114383616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114382528)))]; + tensor module_layers_4_norm_conv_bias = const()[name = tensor("module_layers_4_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114387776)))]; + tensor module_layers_4_norm_conv_weight = const()[name = tensor("module_layers_4_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114391936)))]; + tensor module_layers_4_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(114396096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116495424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116493312)))]; + tensor module_layers_4_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116503680))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116514048))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116512960)))]; + tensor module_layers_4_conv_batch_norm_bias = const()[name = tensor("module_layers_4_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116518208)))]; + tensor module_layers_4_conv_batch_norm_weight = const()[name = tensor("module_layers_4_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116522368)))]; + tensor module_layers_4_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116526528))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117576256))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117575168)))]; + tensor module_layers_4_norm_feed_forward2_bias = const()[name = tensor("module_layers_4_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117580416)))]; + tensor module_layers_4_norm_feed_forward2_weight = const()[name = tensor("module_layers_4_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117584576)))]; + tensor module_layers_4_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117588736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121787264))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121783104)))]; + tensor module_layers_4_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_4_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121803712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125999168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125998080)))]; + tensor module_layers_4_norm_out_bias = const()[name = tensor("module_layers_4_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126003328)))]; + tensor module_layers_4_norm_out_weight = const()[name = tensor("module_layers_4_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126007488)))]; + tensor module_layers_5_norm_feed_forward1_bias = const()[name = tensor("module_layers_5_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126011648)))]; + tensor module_layers_5_norm_feed_forward1_weight = const()[name = tensor("module_layers_5_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126015808)))]; + tensor module_layers_5_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126019968))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130218496))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130214336)))]; + tensor module_layers_5_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130234944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134430400))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134429312)))]; + tensor module_layers_5_norm_self_att_bias = const()[name = tensor("module_layers_5_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134434560)))]; + tensor module_layers_5_norm_self_att_weight = const()[name = tensor("module_layers_5_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134438720)))]; + tensor module_layers_5_self_attn_pos_bias_v = const()[name = tensor("module_layers_5_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134442880)))]; + tensor module_layers_5_self_attn_pos_bias_u = const()[name = tensor("module_layers_5_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134447040)))]; + tensor module_layers_5_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134451200))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135500928))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135499840)))]; + tensor module_layers_5_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135505088))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136554816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136553728)))]; + tensor module_layers_5_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136558976))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137608704))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137607616)))]; + tensor module_layers_5_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(137612864))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138662592))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138661504)))]; + tensor module_layers_5_norm_conv_bias = const()[name = tensor("module_layers_5_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138666752)))]; + tensor module_layers_5_norm_conv_weight = const()[name = tensor("module_layers_5_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138670912)))]; + tensor module_layers_5_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138675072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140774400))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140772288)))]; + tensor module_layers_5_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140782656))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140793024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140791936)))]; + tensor module_layers_5_conv_batch_norm_bias = const()[name = tensor("module_layers_5_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140797184)))]; + tensor module_layers_5_conv_batch_norm_weight = const()[name = tensor("module_layers_5_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140801344)))]; + tensor module_layers_5_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140805504))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141855232))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141854144)))]; + tensor module_layers_5_norm_feed_forward2_bias = const()[name = tensor("module_layers_5_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141859392)))]; + tensor module_layers_5_norm_feed_forward2_weight = const()[name = tensor("module_layers_5_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141863552)))]; + tensor module_layers_5_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141867712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146066240))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146062080)))]; + tensor module_layers_5_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_5_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146082688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150278144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150277056)))]; + tensor module_layers_5_norm_out_bias = const()[name = tensor("module_layers_5_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150282304)))]; + tensor module_layers_5_norm_out_weight = const()[name = tensor("module_layers_5_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150286464)))]; + tensor module_layers_6_norm_feed_forward1_bias = const()[name = tensor("module_layers_6_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150290624)))]; + tensor module_layers_6_norm_feed_forward1_weight = const()[name = tensor("module_layers_6_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150294784)))]; + tensor module_layers_6_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150298944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154497472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154493312)))]; + tensor module_layers_6_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154513920))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158709376))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158708288)))]; + tensor module_layers_6_norm_self_att_bias = const()[name = tensor("module_layers_6_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158713536)))]; + tensor module_layers_6_norm_self_att_weight = const()[name = tensor("module_layers_6_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158717696)))]; + tensor module_layers_6_self_attn_pos_bias_v = const()[name = tensor("module_layers_6_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158721856)))]; + tensor module_layers_6_self_attn_pos_bias_u = const()[name = tensor("module_layers_6_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158726016)))]; + tensor module_layers_6_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158730176))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159779904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159778816)))]; + tensor module_layers_6_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159784064))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160833792))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160832704)))]; + tensor module_layers_6_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160837952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161887680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161886592)))]; + tensor module_layers_6_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161891840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162941568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162940480)))]; + tensor module_layers_6_norm_conv_bias = const()[name = tensor("module_layers_6_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162945728)))]; + tensor module_layers_6_norm_conv_weight = const()[name = tensor("module_layers_6_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162949888)))]; + tensor module_layers_6_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(162954048))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165053376))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165051264)))]; + tensor module_layers_6_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165061632))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165072000))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165070912)))]; + tensor module_layers_6_conv_batch_norm_bias = const()[name = tensor("module_layers_6_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165076160)))]; + tensor module_layers_6_conv_batch_norm_weight = const()[name = tensor("module_layers_6_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165080320)))]; + tensor module_layers_6_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165084480))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166134208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166133120)))]; + tensor module_layers_6_norm_feed_forward2_bias = const()[name = tensor("module_layers_6_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166138368)))]; + tensor module_layers_6_norm_feed_forward2_weight = const()[name = tensor("module_layers_6_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166142528)))]; + tensor module_layers_6_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166146688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170345216))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170341056)))]; + tensor module_layers_6_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_6_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170361664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174557120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174556032)))]; + tensor module_layers_6_norm_out_bias = const()[name = tensor("module_layers_6_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174561280)))]; + tensor module_layers_6_norm_out_weight = const()[name = tensor("module_layers_6_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174565440)))]; + tensor module_layers_7_norm_feed_forward1_bias = const()[name = tensor("module_layers_7_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174569600)))]; + tensor module_layers_7_norm_feed_forward1_weight = const()[name = tensor("module_layers_7_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174573760)))]; + tensor module_layers_7_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174577920))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178776448))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178772288)))]; + tensor module_layers_7_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178792896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182988352))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182987264)))]; + tensor module_layers_7_norm_self_att_bias = const()[name = tensor("module_layers_7_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182992512)))]; + tensor module_layers_7_norm_self_att_weight = const()[name = tensor("module_layers_7_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182996672)))]; + tensor module_layers_7_self_attn_pos_bias_v = const()[name = tensor("module_layers_7_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183000832)))]; + tensor module_layers_7_self_attn_pos_bias_u = const()[name = tensor("module_layers_7_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183004992)))]; + tensor module_layers_7_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183009152))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184058880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184057792)))]; + tensor module_layers_7_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184063040))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185112768))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185111680)))]; + tensor module_layers_7_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185116928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186166656))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186165568)))]; + tensor module_layers_7_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186170816))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187220544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187219456)))]; + tensor module_layers_7_norm_conv_bias = const()[name = tensor("module_layers_7_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187224704)))]; + tensor module_layers_7_norm_conv_weight = const()[name = tensor("module_layers_7_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187228864)))]; + tensor module_layers_7_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(187233024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189332352))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189330240)))]; + tensor module_layers_7_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189340608))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189350976))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189349888)))]; + tensor module_layers_7_conv_batch_norm_bias = const()[name = tensor("module_layers_7_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189355136)))]; + tensor module_layers_7_conv_batch_norm_weight = const()[name = tensor("module_layers_7_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189359296)))]; + tensor module_layers_7_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(189363456))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190413184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190412096)))]; + tensor module_layers_7_norm_feed_forward2_bias = const()[name = tensor("module_layers_7_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190417344)))]; + tensor module_layers_7_norm_feed_forward2_weight = const()[name = tensor("module_layers_7_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190421504)))]; + tensor module_layers_7_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190425664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194624192))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194620032)))]; + tensor module_layers_7_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_7_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194640640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198836096))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198835008)))]; + tensor module_layers_7_norm_out_bias = const()[name = tensor("module_layers_7_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198840256)))]; + tensor module_layers_7_norm_out_weight = const()[name = tensor("module_layers_7_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198844416)))]; + tensor module_layers_8_norm_feed_forward1_bias = const()[name = tensor("module_layers_8_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198848576)))]; + tensor module_layers_8_norm_feed_forward1_weight = const()[name = tensor("module_layers_8_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198852736)))]; + tensor module_layers_8_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198856896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203055424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203051264)))]; + tensor module_layers_8_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203071872))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207267328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207266240)))]; + tensor module_layers_8_norm_self_att_bias = const()[name = tensor("module_layers_8_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207271488)))]; + tensor module_layers_8_norm_self_att_weight = const()[name = tensor("module_layers_8_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207275648)))]; + tensor module_layers_8_self_attn_pos_bias_v = const()[name = tensor("module_layers_8_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207279808)))]; + tensor module_layers_8_self_attn_pos_bias_u = const()[name = tensor("module_layers_8_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207283968)))]; + tensor module_layers_8_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207288128))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208337856))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208336768)))]; + tensor module_layers_8_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208342016))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209391744))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209390656)))]; + tensor module_layers_8_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209395904))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210445632))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210444544)))]; + tensor module_layers_8_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210449792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211499520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211498432)))]; + tensor module_layers_8_norm_conv_bias = const()[name = tensor("module_layers_8_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211503680)))]; + tensor module_layers_8_norm_conv_weight = const()[name = tensor("module_layers_8_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211507840)))]; + tensor module_layers_8_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211512000))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213611328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213609216)))]; + tensor module_layers_8_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213619584))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213629952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213628864)))]; + tensor module_layers_8_conv_batch_norm_bias = const()[name = tensor("module_layers_8_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213634112)))]; + tensor module_layers_8_conv_batch_norm_weight = const()[name = tensor("module_layers_8_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213638272)))]; + tensor module_layers_8_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213642432))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214692160))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214691072)))]; + tensor module_layers_8_norm_feed_forward2_bias = const()[name = tensor("module_layers_8_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214696320)))]; + tensor module_layers_8_norm_feed_forward2_weight = const()[name = tensor("module_layers_8_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214700480)))]; + tensor module_layers_8_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(214704640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218903168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218899008)))]; + tensor module_layers_8_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_8_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218919616))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(223115072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(223113984)))]; + tensor module_layers_8_norm_out_bias = const()[name = tensor("module_layers_8_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(223119232)))]; + tensor module_layers_8_norm_out_weight = const()[name = tensor("module_layers_8_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(223123392)))]; + tensor module_layers_9_norm_feed_forward1_bias = const()[name = tensor("module_layers_9_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(223127552)))]; + tensor module_layers_9_norm_feed_forward1_weight = const()[name = tensor("module_layers_9_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(223131712)))]; + tensor module_layers_9_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(223135872))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227334400))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227330240)))]; + tensor module_layers_9_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(227350848))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231546304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231545216)))]; + tensor module_layers_9_norm_self_att_bias = const()[name = tensor("module_layers_9_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231550464)))]; + tensor module_layers_9_norm_self_att_weight = const()[name = tensor("module_layers_9_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231554624)))]; + tensor module_layers_9_self_attn_pos_bias_v = const()[name = tensor("module_layers_9_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231558784)))]; + tensor module_layers_9_self_attn_pos_bias_u = const()[name = tensor("module_layers_9_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231562944)))]; + tensor module_layers_9_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231567104))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232616832))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232615744)))]; + tensor module_layers_9_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(232620992))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233670720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233669632)))]; + tensor module_layers_9_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(233674880))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234724608))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234723520)))]; + tensor module_layers_9_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(234728768))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235778496))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235777408)))]; + tensor module_layers_9_norm_conv_bias = const()[name = tensor("module_layers_9_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235782656)))]; + tensor module_layers_9_norm_conv_weight = const()[name = tensor("module_layers_9_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235786816)))]; + tensor module_layers_9_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(235790976))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237890304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237888192)))]; + tensor module_layers_9_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237898560))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237908928))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237907840)))]; + tensor module_layers_9_conv_batch_norm_bias = const()[name = tensor("module_layers_9_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237913088)))]; + tensor module_layers_9_conv_batch_norm_weight = const()[name = tensor("module_layers_9_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237917248)))]; + tensor module_layers_9_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(237921408))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238971136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238970048)))]; + tensor module_layers_9_norm_feed_forward2_bias = const()[name = tensor("module_layers_9_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238975296)))]; + tensor module_layers_9_norm_feed_forward2_weight = const()[name = tensor("module_layers_9_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238979456)))]; + tensor module_layers_9_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(238983616))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(243182144))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(243177984)))]; + tensor module_layers_9_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_9_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(243198592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247394048))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247392960)))]; + tensor module_layers_9_norm_out_bias = const()[name = tensor("module_layers_9_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247398208)))]; + tensor module_layers_9_norm_out_weight = const()[name = tensor("module_layers_9_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247402368)))]; + tensor module_layers_10_norm_feed_forward1_bias = const()[name = tensor("module_layers_10_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247406528)))]; + tensor module_layers_10_norm_feed_forward1_weight = const()[name = tensor("module_layers_10_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247410688)))]; + tensor module_layers_10_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(247414848))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251613376))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251609216)))]; + tensor module_layers_10_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251629824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255825280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255824192)))]; + tensor module_layers_10_norm_self_att_bias = const()[name = tensor("module_layers_10_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255829440)))]; + tensor module_layers_10_norm_self_att_weight = const()[name = tensor("module_layers_10_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255833600)))]; + tensor module_layers_10_self_attn_pos_bias_v = const()[name = tensor("module_layers_10_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255837760)))]; + tensor module_layers_10_self_attn_pos_bias_u = const()[name = tensor("module_layers_10_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255841920)))]; + tensor module_layers_10_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(255846080))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256895808))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256894720)))]; + tensor module_layers_10_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(256899968))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257949696))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257948608)))]; + tensor module_layers_10_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257953856))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259003584))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259002496)))]; + tensor module_layers_10_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(259007744))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260057472))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260056384)))]; + tensor module_layers_10_norm_conv_bias = const()[name = tensor("module_layers_10_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260061632)))]; + tensor module_layers_10_norm_conv_weight = const()[name = tensor("module_layers_10_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260065792)))]; + tensor module_layers_10_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(260069952))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262169280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262167168)))]; + tensor module_layers_10_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262177536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262187904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262186816)))]; + tensor module_layers_10_conv_batch_norm_bias = const()[name = tensor("module_layers_10_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262192064)))]; + tensor module_layers_10_conv_batch_norm_weight = const()[name = tensor("module_layers_10_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262196224)))]; + tensor module_layers_10_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(262200384))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263250112))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263249024)))]; + tensor module_layers_10_norm_feed_forward2_bias = const()[name = tensor("module_layers_10_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263254272)))]; + tensor module_layers_10_norm_feed_forward2_weight = const()[name = tensor("module_layers_10_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263258432)))]; + tensor module_layers_10_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263262592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267461120))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267456960)))]; + tensor module_layers_10_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_10_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(267477568))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271673024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271671936)))]; + tensor module_layers_10_norm_out_bias = const()[name = tensor("module_layers_10_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271677184)))]; + tensor module_layers_10_norm_out_weight = const()[name = tensor("module_layers_10_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271681344)))]; + tensor module_layers_11_norm_feed_forward1_bias = const()[name = tensor("module_layers_11_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271685504)))]; + tensor module_layers_11_norm_feed_forward1_weight = const()[name = tensor("module_layers_11_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271689664)))]; + tensor module_layers_11_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271693824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275892352))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275888192)))]; + tensor module_layers_11_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(275908800))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280104256))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280103168)))]; + tensor module_layers_11_norm_self_att_bias = const()[name = tensor("module_layers_11_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280108416)))]; + tensor module_layers_11_norm_self_att_weight = const()[name = tensor("module_layers_11_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280112576)))]; + tensor module_layers_11_self_attn_pos_bias_v = const()[name = tensor("module_layers_11_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280116736)))]; + tensor module_layers_11_self_attn_pos_bias_u = const()[name = tensor("module_layers_11_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280120896)))]; + tensor module_layers_11_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(280125056))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(281174784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(281173696)))]; + tensor module_layers_11_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(281178944))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282228672))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282227584)))]; + tensor module_layers_11_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(282232832))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283282560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283281472)))]; + tensor module_layers_11_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(283286720))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284336448))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284335360)))]; + tensor module_layers_11_norm_conv_bias = const()[name = tensor("module_layers_11_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284340608)))]; + tensor module_layers_11_norm_conv_weight = const()[name = tensor("module_layers_11_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284344768)))]; + tensor module_layers_11_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284348928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286448256))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286446144)))]; + tensor module_layers_11_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286456512))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286466880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286465792)))]; + tensor module_layers_11_conv_batch_norm_bias = const()[name = tensor("module_layers_11_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286471040)))]; + tensor module_layers_11_conv_batch_norm_weight = const()[name = tensor("module_layers_11_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286475200)))]; + tensor module_layers_11_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(286479360))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287529088))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287528000)))]; + tensor module_layers_11_norm_feed_forward2_bias = const()[name = tensor("module_layers_11_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287533248)))]; + tensor module_layers_11_norm_feed_forward2_weight = const()[name = tensor("module_layers_11_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287537408)))]; + tensor module_layers_11_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287541568))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291740096))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291735936)))]; + tensor module_layers_11_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_11_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(291756544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295952000))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295950912)))]; + tensor module_layers_11_norm_out_bias = const()[name = tensor("module_layers_11_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295956160)))]; + tensor module_layers_11_norm_out_weight = const()[name = tensor("module_layers_11_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295960320)))]; + tensor module_layers_12_norm_feed_forward1_bias = const()[name = tensor("module_layers_12_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295964480)))]; + tensor module_layers_12_norm_feed_forward1_weight = const()[name = tensor("module_layers_12_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295968640)))]; + tensor module_layers_12_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(295972800))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(300171328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(300167168)))]; + tensor module_layers_12_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(300187776))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304383232))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304382144)))]; + tensor module_layers_12_norm_self_att_bias = const()[name = tensor("module_layers_12_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304387392)))]; + tensor module_layers_12_norm_self_att_weight = const()[name = tensor("module_layers_12_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304391552)))]; + tensor module_layers_12_self_attn_pos_bias_v = const()[name = tensor("module_layers_12_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304395712)))]; + tensor module_layers_12_self_attn_pos_bias_u = const()[name = tensor("module_layers_12_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304399872)))]; + tensor module_layers_12_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(304404032))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305453760))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305452672)))]; + tensor module_layers_12_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(305457920))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306507648))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306506560)))]; + tensor module_layers_12_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(306511808))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307561536))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307560448)))]; + tensor module_layers_12_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(307565696))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308615424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308614336)))]; + tensor module_layers_12_norm_conv_bias = const()[name = tensor("module_layers_12_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308619584)))]; + tensor module_layers_12_norm_conv_weight = const()[name = tensor("module_layers_12_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308623744)))]; + tensor module_layers_12_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(308627904))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310727232))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310725120)))]; + tensor module_layers_12_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310735488))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310745856))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310744768)))]; + tensor module_layers_12_conv_batch_norm_bias = const()[name = tensor("module_layers_12_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310750016)))]; + tensor module_layers_12_conv_batch_norm_weight = const()[name = tensor("module_layers_12_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310754176)))]; + tensor module_layers_12_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310758336))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311808064))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311806976)))]; + tensor module_layers_12_norm_feed_forward2_bias = const()[name = tensor("module_layers_12_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311812224)))]; + tensor module_layers_12_norm_feed_forward2_weight = const()[name = tensor("module_layers_12_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311816384)))]; + tensor module_layers_12_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(311820544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316019072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316014912)))]; + tensor module_layers_12_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_12_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(316035520))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320230976))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320229888)))]; + tensor module_layers_12_norm_out_bias = const()[name = tensor("module_layers_12_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320235136)))]; + tensor module_layers_12_norm_out_weight = const()[name = tensor("module_layers_12_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320239296)))]; + tensor module_layers_13_norm_feed_forward1_bias = const()[name = tensor("module_layers_13_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320243456)))]; + tensor module_layers_13_norm_feed_forward1_weight = const()[name = tensor("module_layers_13_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320247616)))]; + tensor module_layers_13_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(320251776))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324450304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324446144)))]; + tensor module_layers_13_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(324466752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328662208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328661120)))]; + tensor module_layers_13_norm_self_att_bias = const()[name = tensor("module_layers_13_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328666368)))]; + tensor module_layers_13_norm_self_att_weight = const()[name = tensor("module_layers_13_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328670528)))]; + tensor module_layers_13_self_attn_pos_bias_v = const()[name = tensor("module_layers_13_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328674688)))]; + tensor module_layers_13_self_attn_pos_bias_u = const()[name = tensor("module_layers_13_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328678848)))]; + tensor module_layers_13_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(328683008))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(329732736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(329731648)))]; + tensor module_layers_13_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(329736896))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330786624))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330785536)))]; + tensor module_layers_13_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330790784))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331840512))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331839424)))]; + tensor module_layers_13_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(331844672))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332894400))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332893312)))]; + tensor module_layers_13_norm_conv_bias = const()[name = tensor("module_layers_13_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332898560)))]; + tensor module_layers_13_norm_conv_weight = const()[name = tensor("module_layers_13_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332902720)))]; + tensor module_layers_13_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(332906880))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335006208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335004096)))]; + tensor module_layers_13_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335014464))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335024832))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335023744)))]; + tensor module_layers_13_conv_batch_norm_bias = const()[name = tensor("module_layers_13_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335028992)))]; + tensor module_layers_13_conv_batch_norm_weight = const()[name = tensor("module_layers_13_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335033152)))]; + tensor module_layers_13_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(335037312))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336087040))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336085952)))]; + tensor module_layers_13_norm_feed_forward2_bias = const()[name = tensor("module_layers_13_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336091200)))]; + tensor module_layers_13_norm_feed_forward2_weight = const()[name = tensor("module_layers_13_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336095360)))]; + tensor module_layers_13_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336099520))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(340298048))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(340293888)))]; + tensor module_layers_13_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_13_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(340314496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344509952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344508864)))]; + tensor module_layers_13_norm_out_bias = const()[name = tensor("module_layers_13_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344514112)))]; + tensor module_layers_13_norm_out_weight = const()[name = tensor("module_layers_13_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344518272)))]; + tensor module_layers_14_norm_feed_forward1_bias = const()[name = tensor("module_layers_14_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344522432)))]; + tensor module_layers_14_norm_feed_forward1_weight = const()[name = tensor("module_layers_14_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344526592)))]; + tensor module_layers_14_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(344530752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348729280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348725120)))]; + tensor module_layers_14_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(348745728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352941184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352940096)))]; + tensor module_layers_14_norm_self_att_bias = const()[name = tensor("module_layers_14_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352945344)))]; + tensor module_layers_14_norm_self_att_weight = const()[name = tensor("module_layers_14_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352949504)))]; + tensor module_layers_14_self_attn_pos_bias_v = const()[name = tensor("module_layers_14_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352953664)))]; + tensor module_layers_14_self_attn_pos_bias_u = const()[name = tensor("module_layers_14_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352957824)))]; + tensor module_layers_14_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(352961984))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(354011712))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(354010624)))]; + tensor module_layers_14_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(354015872))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355065600))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355064512)))]; + tensor module_layers_14_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(355069760))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356119488))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356118400)))]; + tensor module_layers_14_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(356123648))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357173376))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357172288)))]; + tensor module_layers_14_norm_conv_bias = const()[name = tensor("module_layers_14_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357177536)))]; + tensor module_layers_14_norm_conv_weight = const()[name = tensor("module_layers_14_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357181696)))]; + tensor module_layers_14_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(357185856))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359285184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359283072)))]; + tensor module_layers_14_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359293440))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359303808))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359302720)))]; + tensor module_layers_14_conv_batch_norm_bias = const()[name = tensor("module_layers_14_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359307968)))]; + tensor module_layers_14_conv_batch_norm_weight = const()[name = tensor("module_layers_14_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359312128)))]; + tensor module_layers_14_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(359316288))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360366016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360364928)))]; + tensor module_layers_14_norm_feed_forward2_bias = const()[name = tensor("module_layers_14_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360370176)))]; + tensor module_layers_14_norm_feed_forward2_weight = const()[name = tensor("module_layers_14_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360374336)))]; + tensor module_layers_14_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(360378496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364577024))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364572864)))]; + tensor module_layers_14_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_14_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(364593472))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368788928))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368787840)))]; + tensor module_layers_14_norm_out_bias = const()[name = tensor("module_layers_14_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368793088)))]; + tensor module_layers_14_norm_out_weight = const()[name = tensor("module_layers_14_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368797248)))]; + tensor module_layers_15_norm_feed_forward1_bias = const()[name = tensor("module_layers_15_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368801408)))]; + tensor module_layers_15_norm_feed_forward1_weight = const()[name = tensor("module_layers_15_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368805568)))]; + tensor module_layers_15_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(368809728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(373008256))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(373004096)))]; + tensor module_layers_15_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(373024704))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377220160))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377219072)))]; + tensor module_layers_15_norm_self_att_bias = const()[name = tensor("module_layers_15_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377224320)))]; + tensor module_layers_15_norm_self_att_weight = const()[name = tensor("module_layers_15_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377228480)))]; + tensor module_layers_15_self_attn_pos_bias_v = const()[name = tensor("module_layers_15_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377232640)))]; + tensor module_layers_15_self_attn_pos_bias_u = const()[name = tensor("module_layers_15_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377236800)))]; + tensor module_layers_15_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(377240960))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378290688))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378289600)))]; + tensor module_layers_15_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(378294848))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379344576))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379343488)))]; + tensor module_layers_15_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(379348736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380398464))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380397376)))]; + tensor module_layers_15_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(380402624))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381452352))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381451264)))]; + tensor module_layers_15_norm_conv_bias = const()[name = tensor("module_layers_15_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381456512)))]; + tensor module_layers_15_norm_conv_weight = const()[name = tensor("module_layers_15_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381460672)))]; + tensor module_layers_15_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(381464832))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383564160))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383562048)))]; + tensor module_layers_15_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383572416))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383582784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383581696)))]; + tensor module_layers_15_conv_batch_norm_bias = const()[name = tensor("module_layers_15_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383586944)))]; + tensor module_layers_15_conv_batch_norm_weight = const()[name = tensor("module_layers_15_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383591104)))]; + tensor module_layers_15_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(383595264))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384644992))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384643904)))]; + tensor module_layers_15_norm_feed_forward2_bias = const()[name = tensor("module_layers_15_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384649152)))]; + tensor module_layers_15_norm_feed_forward2_weight = const()[name = tensor("module_layers_15_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384653312)))]; + tensor module_layers_15_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(384657472))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388856000))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388851840)))]; + tensor module_layers_15_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_15_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(388872448))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393067904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393066816)))]; + tensor module_layers_15_norm_out_bias = const()[name = tensor("module_layers_15_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393072064)))]; + tensor module_layers_15_norm_out_weight = const()[name = tensor("module_layers_15_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393076224)))]; + tensor module_layers_16_norm_feed_forward1_bias = const()[name = tensor("module_layers_16_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393080384)))]; + tensor module_layers_16_norm_feed_forward1_weight = const()[name = tensor("module_layers_16_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393084544)))]; + tensor module_layers_16_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(393088704))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397287232))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397283072)))]; + tensor module_layers_16_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(397303680))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401499136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401498048)))]; + tensor module_layers_16_norm_self_att_bias = const()[name = tensor("module_layers_16_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401503296)))]; + tensor module_layers_16_norm_self_att_weight = const()[name = tensor("module_layers_16_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401507456)))]; + tensor module_layers_16_self_attn_pos_bias_v = const()[name = tensor("module_layers_16_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401511616)))]; + tensor module_layers_16_self_attn_pos_bias_u = const()[name = tensor("module_layers_16_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401515776)))]; + tensor module_layers_16_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(401519936))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402569664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402568576)))]; + tensor module_layers_16_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402573824))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403623552))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403622464)))]; + tensor module_layers_16_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(403627712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(404677440))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(404676352)))]; + tensor module_layers_16_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(404681600))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405731328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405730240)))]; + tensor module_layers_16_norm_conv_bias = const()[name = tensor("module_layers_16_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405735488)))]; + tensor module_layers_16_norm_conv_weight = const()[name = tensor("module_layers_16_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405739648)))]; + tensor module_layers_16_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405743808))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407843136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407841024)))]; + tensor module_layers_16_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407851392))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407861760))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407860672)))]; + tensor module_layers_16_conv_batch_norm_bias = const()[name = tensor("module_layers_16_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407865920)))]; + tensor module_layers_16_conv_batch_norm_weight = const()[name = tensor("module_layers_16_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407870080)))]; + tensor module_layers_16_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(407874240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408923968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408922880)))]; + tensor module_layers_16_norm_feed_forward2_bias = const()[name = tensor("module_layers_16_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408928128)))]; + tensor module_layers_16_norm_feed_forward2_weight = const()[name = tensor("module_layers_16_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408932288)))]; + tensor module_layers_16_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408936448))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413134976))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413130816)))]; + tensor module_layers_16_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_16_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(413151424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417346880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417345792)))]; + tensor module_layers_16_norm_out_bias = const()[name = tensor("module_layers_16_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417351040)))]; + tensor module_layers_16_norm_out_weight = const()[name = tensor("module_layers_16_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417355200)))]; + tensor module_layers_17_norm_feed_forward1_bias = const()[name = tensor("module_layers_17_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417359360)))]; + tensor module_layers_17_norm_feed_forward1_weight = const()[name = tensor("module_layers_17_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417363520)))]; + tensor module_layers_17_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(417367680))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421566208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421562048)))]; + tensor module_layers_17_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(421582656))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425778112))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425777024)))]; + tensor module_layers_17_norm_self_att_bias = const()[name = tensor("module_layers_17_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425782272)))]; + tensor module_layers_17_norm_self_att_weight = const()[name = tensor("module_layers_17_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425786432)))]; + tensor module_layers_17_self_attn_pos_bias_v = const()[name = tensor("module_layers_17_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425790592)))]; + tensor module_layers_17_self_attn_pos_bias_u = const()[name = tensor("module_layers_17_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425794752)))]; + tensor module_layers_17_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(425798912))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426848640))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426847552)))]; + tensor module_layers_17_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(426852800))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427902528))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427901440)))]; + tensor module_layers_17_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(427906688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428956416))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428955328)))]; + tensor module_layers_17_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428960576))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430010304))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430009216)))]; + tensor module_layers_17_norm_conv_bias = const()[name = tensor("module_layers_17_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430014464)))]; + tensor module_layers_17_norm_conv_weight = const()[name = tensor("module_layers_17_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430018624)))]; + tensor module_layers_17_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(430022784))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432122112))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432120000)))]; + tensor module_layers_17_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432130368))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432140736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432139648)))]; + tensor module_layers_17_conv_batch_norm_bias = const()[name = tensor("module_layers_17_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432144896)))]; + tensor module_layers_17_conv_batch_norm_weight = const()[name = tensor("module_layers_17_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432149056)))]; + tensor module_layers_17_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(432153216))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433202944))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433201856)))]; + tensor module_layers_17_norm_feed_forward2_bias = const()[name = tensor("module_layers_17_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433207104)))]; + tensor module_layers_17_norm_feed_forward2_weight = const()[name = tensor("module_layers_17_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433211264)))]; + tensor module_layers_17_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(433215424))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437413952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437409792)))]; + tensor module_layers_17_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_17_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(437430400))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441625856))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441624768)))]; + tensor module_layers_17_norm_out_bias = const()[name = tensor("module_layers_17_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441630016)))]; + tensor module_layers_17_norm_out_weight = const()[name = tensor("module_layers_17_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441634176)))]; + tensor module_layers_18_norm_feed_forward1_bias = const()[name = tensor("module_layers_18_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441638336)))]; + tensor module_layers_18_norm_feed_forward1_weight = const()[name = tensor("module_layers_18_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441642496)))]; + tensor module_layers_18_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441646656))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445845184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445841024)))]; + tensor module_layers_18_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(445861632))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450057088))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450056000)))]; + tensor module_layers_18_norm_self_att_bias = const()[name = tensor("module_layers_18_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450061248)))]; + tensor module_layers_18_norm_self_att_weight = const()[name = tensor("module_layers_18_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450065408)))]; + tensor module_layers_18_self_attn_pos_bias_v = const()[name = tensor("module_layers_18_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450069568)))]; + tensor module_layers_18_self_attn_pos_bias_u = const()[name = tensor("module_layers_18_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450073728)))]; + tensor module_layers_18_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(450077888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451127616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451126528)))]; + tensor module_layers_18_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451131776))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452181504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452180416)))]; + tensor module_layers_18_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(452185664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(453235392))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(453234304)))]; + tensor module_layers_18_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(453239552))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454289280))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454288192)))]; + tensor module_layers_18_norm_conv_bias = const()[name = tensor("module_layers_18_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454293440)))]; + tensor module_layers_18_norm_conv_weight = const()[name = tensor("module_layers_18_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454297600)))]; + tensor module_layers_18_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454301760))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456401088))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456398976)))]; + tensor module_layers_18_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456409344))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456419712))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456418624)))]; + tensor module_layers_18_conv_batch_norm_bias = const()[name = tensor("module_layers_18_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456423872)))]; + tensor module_layers_18_conv_batch_norm_weight = const()[name = tensor("module_layers_18_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456428032)))]; + tensor module_layers_18_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(456432192))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457481920))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457480832)))]; + tensor module_layers_18_norm_feed_forward2_bias = const()[name = tensor("module_layers_18_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457486080)))]; + tensor module_layers_18_norm_feed_forward2_weight = const()[name = tensor("module_layers_18_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457490240)))]; + tensor module_layers_18_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(457494400))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461692928))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461688768)))]; + tensor module_layers_18_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_18_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(461709376))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(465904832))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(465903744)))]; + tensor module_layers_18_norm_out_bias = const()[name = tensor("module_layers_18_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(465908992)))]; + tensor module_layers_18_norm_out_weight = const()[name = tensor("module_layers_18_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(465913152)))]; + tensor module_layers_19_norm_feed_forward1_bias = const()[name = tensor("module_layers_19_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(465917312)))]; + tensor module_layers_19_norm_feed_forward1_weight = const()[name = tensor("module_layers_19_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(465921472)))]; + tensor module_layers_19_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(465925632))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470124160))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470120000)))]; + tensor module_layers_19_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(470140608))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474336064))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474334976)))]; + tensor module_layers_19_norm_self_att_bias = const()[name = tensor("module_layers_19_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474340224)))]; + tensor module_layers_19_norm_self_att_weight = const()[name = tensor("module_layers_19_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474344384)))]; + tensor module_layers_19_self_attn_pos_bias_v = const()[name = tensor("module_layers_19_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474348544)))]; + tensor module_layers_19_self_attn_pos_bias_u = const()[name = tensor("module_layers_19_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474352704)))]; + tensor module_layers_19_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(474356864))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(475406592))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(475405504)))]; + tensor module_layers_19_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(475410752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476460480))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476459392)))]; + tensor module_layers_19_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(476464640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(477514368))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(477513280)))]; + tensor module_layers_19_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(477518528))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478568256))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478567168)))]; + tensor module_layers_19_norm_conv_bias = const()[name = tensor("module_layers_19_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478572416)))]; + tensor module_layers_19_norm_conv_weight = const()[name = tensor("module_layers_19_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478576576)))]; + tensor module_layers_19_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(478580736))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480680064))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480677952)))]; + tensor module_layers_19_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480688320))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480698688))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480697600)))]; + tensor module_layers_19_conv_batch_norm_bias = const()[name = tensor("module_layers_19_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480702848)))]; + tensor module_layers_19_conv_batch_norm_weight = const()[name = tensor("module_layers_19_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480707008)))]; + tensor module_layers_19_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480711168))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481760896))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481759808)))]; + tensor module_layers_19_norm_feed_forward2_bias = const()[name = tensor("module_layers_19_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481765056)))]; + tensor module_layers_19_norm_feed_forward2_weight = const()[name = tensor("module_layers_19_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481769216)))]; + tensor module_layers_19_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(481773376))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485971904))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485967744)))]; + tensor module_layers_19_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_19_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(485988352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490183808))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490182720)))]; + tensor module_layers_19_norm_out_bias = const()[name = tensor("module_layers_19_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490187968)))]; + tensor module_layers_19_norm_out_weight = const()[name = tensor("module_layers_19_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490192128)))]; + tensor module_layers_20_norm_feed_forward1_bias = const()[name = tensor("module_layers_20_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490196288)))]; + tensor module_layers_20_norm_feed_forward1_weight = const()[name = tensor("module_layers_20_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490200448)))]; + tensor module_layers_20_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490204608))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494403136))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494398976)))]; + tensor module_layers_20_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494419584))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498615040))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498613952)))]; + tensor module_layers_20_norm_self_att_bias = const()[name = tensor("module_layers_20_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498619200)))]; + tensor module_layers_20_norm_self_att_weight = const()[name = tensor("module_layers_20_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498623360)))]; + tensor module_layers_20_self_attn_pos_bias_v = const()[name = tensor("module_layers_20_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498627520)))]; + tensor module_layers_20_self_attn_pos_bias_u = const()[name = tensor("module_layers_20_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498631680)))]; + tensor module_layers_20_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(498635840))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499685568))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499684480)))]; + tensor module_layers_20_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(499689728))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500739456))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500738368)))]; + tensor module_layers_20_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(500743616))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501793344))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501792256)))]; + tensor module_layers_20_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(501797504))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(502847232))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(502846144)))]; + tensor module_layers_20_norm_conv_bias = const()[name = tensor("module_layers_20_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(502851392)))]; + tensor module_layers_20_norm_conv_weight = const()[name = tensor("module_layers_20_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(502855552)))]; + tensor module_layers_20_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(502859712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504959040))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504956928)))]; + tensor module_layers_20_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504967296))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504977664))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504976576)))]; + tensor module_layers_20_conv_batch_norm_bias = const()[name = tensor("module_layers_20_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504981824)))]; + tensor module_layers_20_conv_batch_norm_weight = const()[name = tensor("module_layers_20_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504985984)))]; + tensor module_layers_20_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(504990144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506039872))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506038784)))]; + tensor module_layers_20_norm_feed_forward2_bias = const()[name = tensor("module_layers_20_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506044032)))]; + tensor module_layers_20_norm_feed_forward2_weight = const()[name = tensor("module_layers_20_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506048192)))]; + tensor module_layers_20_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(506052352))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510250880))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510246720)))]; + tensor module_layers_20_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_20_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(510267328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514462784))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514461696)))]; + tensor module_layers_20_norm_out_bias = const()[name = tensor("module_layers_20_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514466944)))]; + tensor module_layers_20_norm_out_weight = const()[name = tensor("module_layers_20_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514471104)))]; + tensor module_layers_21_norm_feed_forward1_bias = const()[name = tensor("module_layers_21_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514475264)))]; + tensor module_layers_21_norm_feed_forward1_weight = const()[name = tensor("module_layers_21_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514479424)))]; + tensor module_layers_21_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(514483584))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(518682112))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(518677952)))]; + tensor module_layers_21_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(518698560))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522894016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522892928)))]; + tensor module_layers_21_norm_self_att_bias = const()[name = tensor("module_layers_21_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522898176)))]; + tensor module_layers_21_norm_self_att_weight = const()[name = tensor("module_layers_21_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522902336)))]; + tensor module_layers_21_self_attn_pos_bias_v = const()[name = tensor("module_layers_21_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522906496)))]; + tensor module_layers_21_self_attn_pos_bias_u = const()[name = tensor("module_layers_21_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522910656)))]; + tensor module_layers_21_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(522914816))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523964544))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523963456)))]; + tensor module_layers_21_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523968704))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525018432))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525017344)))]; + tensor module_layers_21_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(525022592))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526072320))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526071232)))]; + tensor module_layers_21_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526076480))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527126208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527125120)))]; + tensor module_layers_21_norm_conv_bias = const()[name = tensor("module_layers_21_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527130368)))]; + tensor module_layers_21_norm_conv_weight = const()[name = tensor("module_layers_21_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527134528)))]; + tensor module_layers_21_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(527138688))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529238016))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529235904)))]; + tensor module_layers_21_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529246272))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529256640))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529255552)))]; + tensor module_layers_21_conv_batch_norm_bias = const()[name = tensor("module_layers_21_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529260800)))]; + tensor module_layers_21_conv_batch_norm_weight = const()[name = tensor("module_layers_21_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529264960)))]; + tensor module_layers_21_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(529269120))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530318848))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530317760)))]; + tensor module_layers_21_norm_feed_forward2_bias = const()[name = tensor("module_layers_21_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530323008)))]; + tensor module_layers_21_norm_feed_forward2_weight = const()[name = tensor("module_layers_21_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530327168)))]; + tensor module_layers_21_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530331328))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(534529856))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(534525696)))]; + tensor module_layers_21_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_21_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(534546304))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538741760))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538740672)))]; + tensor module_layers_21_norm_out_bias = const()[name = tensor("module_layers_21_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538745920)))]; + tensor module_layers_21_norm_out_weight = const()[name = tensor("module_layers_21_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538750080)))]; + tensor module_layers_22_norm_feed_forward1_bias = const()[name = tensor("module_layers_22_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538754240)))]; + tensor module_layers_22_norm_feed_forward1_weight = const()[name = tensor("module_layers_22_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538758400)))]; + tensor module_layers_22_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(538762560))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542961088))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542956928)))]; + tensor module_layers_22_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(542977536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547172992))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547171904)))]; + tensor module_layers_22_norm_self_att_bias = const()[name = tensor("module_layers_22_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547177152)))]; + tensor module_layers_22_norm_self_att_weight = const()[name = tensor("module_layers_22_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547181312)))]; + tensor module_layers_22_self_attn_pos_bias_v = const()[name = tensor("module_layers_22_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547185472)))]; + tensor module_layers_22_self_attn_pos_bias_u = const()[name = tensor("module_layers_22_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547189632)))]; + tensor module_layers_22_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(547193792))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548243520))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548242432)))]; + tensor module_layers_22_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(548247680))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549297408))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549296320)))]; + tensor module_layers_22_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(549301568))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550351296))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550350208)))]; + tensor module_layers_22_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(550355456))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551405184))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551404096)))]; + tensor module_layers_22_norm_conv_bias = const()[name = tensor("module_layers_22_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551409344)))]; + tensor module_layers_22_norm_conv_weight = const()[name = tensor("module_layers_22_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551413504)))]; + tensor module_layers_22_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(551417664))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553516992))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553514880)))]; + tensor module_layers_22_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553525248))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553535616))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553534528)))]; + tensor module_layers_22_conv_batch_norm_bias = const()[name = tensor("module_layers_22_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553539776)))]; + tensor module_layers_22_conv_batch_norm_weight = const()[name = tensor("module_layers_22_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553543936)))]; + tensor module_layers_22_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(553548096))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554597824))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554596736)))]; + tensor module_layers_22_norm_feed_forward2_bias = const()[name = tensor("module_layers_22_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554601984)))]; + tensor module_layers_22_norm_feed_forward2_weight = const()[name = tensor("module_layers_22_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554606144)))]; + tensor module_layers_22_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(554610304))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558808832))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558804672)))]; + tensor module_layers_22_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_22_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(558825280))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563020736))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563019648)))]; + tensor module_layers_22_norm_out_bias = const()[name = tensor("module_layers_22_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563024896)))]; + tensor module_layers_22_norm_out_weight = const()[name = tensor("module_layers_22_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563029056)))]; + tensor module_layers_23_norm_feed_forward1_bias = const()[name = tensor("module_layers_23_norm_feed_forward1_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563033216)))]; + tensor module_layers_23_norm_feed_forward1_weight = const()[name = tensor("module_layers_23_norm_feed_forward1_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563037376)))]; + tensor module_layers_23_feed_forward1_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_feed_forward1_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(563041536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567240064))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567235904)))]; + tensor module_layers_23_feed_forward1_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_feed_forward1_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567256512))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571451968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571450880)))]; + tensor module_layers_23_norm_self_att_bias = const()[name = tensor("module_layers_23_norm_self_att_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571456128)))]; + tensor module_layers_23_norm_self_att_weight = const()[name = tensor("module_layers_23_norm_self_att_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571460288)))]; + tensor module_layers_23_self_attn_pos_bias_v = const()[name = tensor("module_layers_23_self_attn_pos_bias_v"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571464448)))]; + tensor module_layers_23_self_attn_pos_bias_u = const()[name = tensor("module_layers_23_self_attn_pos_bias_u"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571468608)))]; + tensor module_layers_23_self_attn_linear_q_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_self_attn_linear_q_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(571472768))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572522496))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572521408)))]; + tensor module_layers_23_self_attn_linear_k_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_self_attn_linear_k_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572526656))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573576384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573575296)))]; + tensor module_layers_23_self_attn_linear_v_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_self_attn_linear_v_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(573580544))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574630272))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574629184)))]; + tensor module_layers_23_self_attn_linear_out_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_self_attn_linear_out_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(574634432))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(575684160))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(575683072)))]; + tensor module_layers_23_norm_conv_bias = const()[name = tensor("module_layers_23_norm_conv_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(575688320)))]; + tensor module_layers_23_norm_conv_weight = const()[name = tensor("module_layers_23_norm_conv_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(575692480)))]; + tensor module_layers_23_conv_pointwise_conv1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_conv_pointwise_conv1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(575696640))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577795968))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577793856)))]; + tensor module_layers_23_conv_depthwise_conv_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_conv_depthwise_conv_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577804224))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577814592))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577813504)))]; + tensor module_layers_23_conv_batch_norm_bias = const()[name = tensor("module_layers_23_conv_batch_norm_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577818752)))]; + tensor module_layers_23_conv_batch_norm_weight = const()[name = tensor("module_layers_23_conv_batch_norm_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577822912)))]; + tensor module_layers_23_conv_pointwise_conv2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_conv_pointwise_conv2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(577827072))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578876800))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578875712)))]; + tensor module_layers_23_norm_feed_forward2_bias = const()[name = tensor("module_layers_23_norm_feed_forward2_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578880960)))]; + tensor module_layers_23_norm_feed_forward2_weight = const()[name = tensor("module_layers_23_norm_feed_forward2_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578885120)))]; + tensor module_layers_23_feed_forward2_linear1_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_feed_forward2_linear1_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(578889280))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(583087808))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(583083648)))]; + tensor module_layers_23_feed_forward2_linear2_weight_quantized = constexpr_affine_dequantize()[axis = tensor(0), name = tensor("module_layers_23_feed_forward2_linear2_weight_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(583104256))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587299712))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587298624)))]; + tensor module_layers_23_norm_out_bias = const()[name = tensor("module_layers_23_norm_out_bias"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587303872)))]; + tensor module_layers_23_norm_out_weight = const()[name = tensor("module_layers_23_norm_out_weight"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587308032)))]; + tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([1, 0, 2, 3])]; + tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([1, 0, 2, 3])]; + tensor var_38 = const()[name = tensor("op_38"), val = tensor(0x1.4f8b58p-17)]; + tensor var_40 = const()[name = tensor("op_40"), val = tensor(0x0p+0)]; + tensor var_41 = const()[name = tensor("op_41"), val = tensor(-0x1.388p+13)]; + tensor var_56 = const()[name = tensor("op_56"), val = tensor(-1)]; + tensor var_65 = const()[name = tensor("op_65"), val = tensor(1)]; + tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; + tensor tensor_1_axes_0 = const()[name = tensor("tensor_1_axes_0"), val = tensor([1])]; + tensor x_1 = transpose(perm = x_1_perm_0, x = mel)[name = tensor("transpose_367")]; + tensor tensor_1 = expand_dims(axes = tensor_1_axes_0, x = x_1)[name = tensor("tensor_1")]; + tensor current_lengths_1_dtype_0 = const()[name = tensor("current_lengths_1_dtype_0"), val = tensor("fp32")]; + tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24]])]; + tensor var_134_axes_0 = const()[name = tensor("op_134_axes_0"), val = tensor([1])]; + tensor var_134 = expand_dims(axes = var_134_axes_0, x = mel_length)[name = tensor("op_134")]; + tensor time_mask_1 = less(x = expand_dims_0, y = var_134)[name = tensor("time_mask_1")]; + tensor var_136_axes_0 = const()[name = tensor("op_136_axes_0"), val = tensor([-1])]; + tensor var_136 = expand_dims(axes = var_136_axes_0, x = time_mask_1)[name = tensor("op_136")]; + tensor var_138_reps_0 = const()[name = tensor("op_138_reps_0"), val = tensor([1, 1, 128])]; + tensor var_138 = tile(reps = var_138_reps_0, x = var_136)[name = tensor("op_138")]; + tensor mask_1_dtype_0 = const()[name = tensor("mask_1_dtype_0"), val = tensor("fp32")]; + tensor var_144_axes_0 = const()[name = tensor("op_144_axes_0"), val = tensor([1])]; + tensor mask_1 = cast(dtype = mask_1_dtype_0, x = var_138)[name = tensor("cast_11")]; + tensor var_144 = expand_dims(axes = var_144_axes_0, x = mask_1)[name = tensor("op_144")]; + tensor input_1 = mul(x = tensor_1, y = var_144)[name = tensor("input_1")]; + tensor const_9 = const()[name = tensor("const_9"), val = tensor(0x0p+0)]; + tensor input_3_pad_0 = const()[name = tensor("input_3_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_3_mode_0 = const()[name = tensor("input_3_mode_0"), val = tensor("constant")]; + tensor input_3 = pad(constant_val = const_9, mode = input_3_mode_0, pad = input_3_pad_0, x = input_1)[name = tensor("input_3")]; + tensor tensor_3_pad_type_0 = const()[name = tensor("tensor_3_pad_type_0"), val = tensor("valid")]; + tensor tensor_3_strides_0 = const()[name = tensor("tensor_3_strides_0"), val = tensor([2, 2])]; + tensor tensor_3_pad_0 = const()[name = tensor("tensor_3_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_3_dilations_0 = const()[name = tensor("tensor_3_dilations_0"), val = tensor([1, 1])]; + tensor tensor_3_groups_0 = const()[name = tensor("tensor_3_groups_0"), val = tensor(1)]; + tensor tensor_3 = conv(bias = module_pre_encode_conv_0_bias, dilations = tensor_3_dilations_0, groups = tensor_3_groups_0, pad = tensor_3_pad_0, pad_type = tensor_3_pad_type_0, strides = tensor_3_strides_0, weight = module_pre_encode_conv_0_weight_quantized, x = input_3)[name = tensor("tensor_3")]; + tensor var_157_promoted = const()[name = tensor("op_157_promoted"), val = tensor(0x1p+1)]; + tensor current_lengths_1 = cast(dtype = current_lengths_1_dtype_0, x = mel_length)[name = tensor("cast_10")]; + tensor var_158 = add(x = current_lengths_1, y = var_157_promoted)[name = tensor("op_158")]; + tensor var_159_promoted = const()[name = tensor("op_159_promoted"), val = tensor(0x1p+0)]; + tensor var_160 = add(x = var_158, y = var_159_promoted)[name = tensor("op_160")]; + tensor var_161_promoted = const()[name = tensor("op_161_promoted"), val = tensor(0x1.8p+1)]; + tensor var_162 = sub(x = var_160, y = var_161_promoted)[name = tensor("op_162")]; + tensor var_53_promoted = const()[name = tensor("op_53_promoted"), val = tensor(0x1p+1)]; + tensor floor_div_0 = floor_div(x = var_162, y = var_53_promoted)[name = tensor("floor_div_0")]; + tensor var_164_promoted = const()[name = tensor("op_164_promoted"), val = tensor(0x1p+0)]; + tensor current_lengths_3 = add(x = floor_div_0, y = var_164_promoted)[name = tensor("current_lengths_3")]; + tensor lengths_19_dtype_0 = const()[name = tensor("lengths_19_dtype_0"), val = tensor("int32")]; + tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12]])]; + tensor var_173_axes_0 = const()[name = tensor("op_173_axes_0"), val = tensor([1])]; + tensor lengths_19 = cast(dtype = lengths_19_dtype_0, x = current_lengths_3)[name = tensor("cast_9")]; + tensor var_173 = expand_dims(axes = var_173_axes_0, x = lengths_19)[name = tensor("op_173")]; + tensor time_mask_3 = less(x = expand_dims_1, y = var_173)[name = tensor("time_mask_3")]; + tensor var_175_axes_0 = const()[name = tensor("op_175_axes_0"), val = tensor([-1])]; + tensor var_175 = expand_dims(axes = var_175_axes_0, x = time_mask_3)[name = tensor("op_175")]; + tensor var_177_reps_0 = const()[name = tensor("op_177_reps_0"), val = tensor([1, 1, 65])]; + tensor var_177 = tile(reps = var_177_reps_0, x = var_175)[name = tensor("op_177")]; + tensor mask_3_dtype_0 = const()[name = tensor("mask_3_dtype_0"), val = tensor("fp32")]; + tensor var_183_axes_0 = const()[name = tensor("op_183_axes_0"), val = tensor([1])]; + tensor mask_3 = cast(dtype = mask_3_dtype_0, x = var_177)[name = tensor("cast_8")]; + tensor var_183 = expand_dims(axes = var_183_axes_0, x = mask_3)[name = tensor("op_183")]; + tensor expanded_mask_3_reps_0 = const()[name = tensor("expanded_mask_3_reps_0"), val = tensor([1, 256, 1, 1])]; + tensor expanded_mask_3 = tile(reps = expanded_mask_3_reps_0, x = var_183)[name = tensor("expanded_mask_3")]; + tensor input_5 = mul(x = tensor_3, y = expanded_mask_3)[name = tensor("input_5")]; + tensor tensor_5 = relu(x = input_5)[name = tensor("tensor_5")]; + tensor input_7 = mul(x = tensor_5, y = expanded_mask_3)[name = tensor("input_7")]; + tensor const_23 = const()[name = tensor("const_23"), val = tensor(0x0p+0)]; + tensor input_9_pad_0 = const()[name = tensor("input_9_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_9_mode_0 = const()[name = tensor("input_9_mode_0"), val = tensor("constant")]; + tensor input_9 = pad(constant_val = const_23, mode = input_9_mode_0, pad = input_9_pad_0, x = input_7)[name = tensor("input_9")]; + tensor tensor_7_pad_type_0 = const()[name = tensor("tensor_7_pad_type_0"), val = tensor("valid")]; + tensor tensor_7_strides_0 = const()[name = tensor("tensor_7_strides_0"), val = tensor([2, 2])]; + tensor tensor_7_groups_0 = const()[name = tensor("tensor_7_groups_0"), val = tensor(256)]; + tensor tensor_7_pad_0 = const()[name = tensor("tensor_7_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_7_dilations_0 = const()[name = tensor("tensor_7_dilations_0"), val = tensor([1, 1])]; + tensor tensor_7 = conv(bias = module_pre_encode_conv_2_bias, dilations = tensor_7_dilations_0, groups = tensor_7_groups_0, pad = tensor_7_pad_0, pad_type = tensor_7_pad_type_0, strides = tensor_7_strides_0, weight = module_pre_encode_conv_2_weight_quantized, x = input_9)[name = tensor("tensor_7")]; + tensor var_205_promoted = const()[name = tensor("op_205_promoted"), val = tensor(0x1p+1)]; + tensor var_206 = add(x = current_lengths_3, y = var_205_promoted)[name = tensor("op_206")]; + tensor var_207_promoted = const()[name = tensor("op_207_promoted"), val = tensor(0x1p+0)]; + tensor var_208 = add(x = var_206, y = var_207_promoted)[name = tensor("op_208")]; + tensor var_209_promoted = const()[name = tensor("op_209_promoted"), val = tensor(0x1.8p+1)]; + tensor var_210 = sub(x = var_208, y = var_209_promoted)[name = tensor("op_210")]; + tensor var_53_promoted_1 = const()[name = tensor("op_53_promoted_1"), val = tensor(0x1p+1)]; + tensor floor_div_1 = floor_div(x = var_210, y = var_53_promoted_1)[name = tensor("floor_div_1")]; + tensor var_212_promoted = const()[name = tensor("op_212_promoted"), val = tensor(0x1p+0)]; + tensor current_lengths_5 = add(x = floor_div_1, y = var_212_promoted)[name = tensor("current_lengths_5")]; + tensor lengths_21_dtype_0 = const()[name = tensor("lengths_21_dtype_0"), val = tensor("int32")]; + tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor([[0, 1, 2, 3, 4, 5, 6]])]; + tensor var_221_axes_0 = const()[name = tensor("op_221_axes_0"), val = tensor([1])]; + tensor lengths_21 = cast(dtype = lengths_21_dtype_0, x = current_lengths_5)[name = tensor("cast_7")]; + tensor var_221 = expand_dims(axes = var_221_axes_0, x = lengths_21)[name = tensor("op_221")]; + tensor time_mask_5 = less(x = expand_dims_2, y = var_221)[name = tensor("time_mask_5")]; + tensor var_223_axes_0 = const()[name = tensor("op_223_axes_0"), val = tensor([-1])]; + tensor var_223 = expand_dims(axes = var_223_axes_0, x = time_mask_5)[name = tensor("op_223")]; + tensor var_225_reps_0 = const()[name = tensor("op_225_reps_0"), val = tensor([1, 1, 33])]; + tensor var_225 = tile(reps = var_225_reps_0, x = var_223)[name = tensor("op_225")]; + tensor mask_5_dtype_0 = const()[name = tensor("mask_5_dtype_0"), val = tensor("fp32")]; + tensor var_231_axes_0 = const()[name = tensor("op_231_axes_0"), val = tensor([1])]; + tensor mask_5 = cast(dtype = mask_5_dtype_0, x = var_225)[name = tensor("cast_6")]; + tensor var_231 = expand_dims(axes = var_231_axes_0, x = mask_5)[name = tensor("op_231")]; + tensor expanded_mask_7_reps_0 = const()[name = tensor("expanded_mask_7_reps_0"), val = tensor([1, 256, 1, 1])]; + tensor expanded_mask_7 = tile(reps = expanded_mask_7_reps_0, x = var_231)[name = tensor("expanded_mask_7")]; + tensor input_11 = mul(x = tensor_7, y = expanded_mask_7)[name = tensor("input_11")]; + tensor tensor_9_pad_type_0 = const()[name = tensor("tensor_9_pad_type_0"), val = tensor("valid")]; + tensor tensor_9_strides_0 = const()[name = tensor("tensor_9_strides_0"), val = tensor([1, 1])]; + tensor tensor_9_pad_0 = const()[name = tensor("tensor_9_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_9_dilations_0 = const()[name = tensor("tensor_9_dilations_0"), val = tensor([1, 1])]; + tensor tensor_9_groups_0 = const()[name = tensor("tensor_9_groups_0"), val = tensor(1)]; + tensor tensor_9 = conv(bias = module_pre_encode_conv_3_bias, dilations = tensor_9_dilations_0, groups = tensor_9_groups_0, pad = tensor_9_pad_0, pad_type = tensor_9_pad_type_0, strides = tensor_9_strides_0, weight = module_pre_encode_conv_3_weight_quantized, x = input_11)[name = tensor("tensor_9")]; + tensor input_13 = mul(x = tensor_9, y = expanded_mask_7)[name = tensor("input_13")]; + tensor tensor_11 = relu(x = input_13)[name = tensor("tensor_11")]; + tensor input_15 = mul(x = tensor_11, y = expanded_mask_7)[name = tensor("input_15")]; + tensor const_41 = const()[name = tensor("const_41"), val = tensor(0x0p+0)]; + tensor input_17_pad_0 = const()[name = tensor("input_17_pad_0"), val = tensor([0, 0, 0, 0, 2, 1, 2, 1])]; + tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("constant")]; + tensor input_17 = pad(constant_val = const_41, mode = input_17_mode_0, pad = input_17_pad_0, x = input_15)[name = tensor("input_17")]; + tensor tensor_13_pad_type_0 = const()[name = tensor("tensor_13_pad_type_0"), val = tensor("valid")]; + tensor tensor_13_strides_0 = const()[name = tensor("tensor_13_strides_0"), val = tensor([2, 2])]; + tensor tensor_13_groups_0 = const()[name = tensor("tensor_13_groups_0"), val = tensor(256)]; + tensor tensor_13_pad_0 = const()[name = tensor("tensor_13_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_13_dilations_0 = const()[name = tensor("tensor_13_dilations_0"), val = tensor([1, 1])]; + tensor tensor_13 = conv(bias = module_pre_encode_conv_5_bias, dilations = tensor_13_dilations_0, groups = tensor_13_groups_0, pad = tensor_13_pad_0, pad_type = tensor_13_pad_type_0, strides = tensor_13_strides_0, weight = module_pre_encode_conv_5_weight_quantized, x = input_17)[name = tensor("tensor_13")]; + tensor var_268_promoted = const()[name = tensor("op_268_promoted"), val = tensor(0x1p+1)]; + tensor var_269 = add(x = current_lengths_5, y = var_268_promoted)[name = tensor("op_269")]; + tensor var_270_promoted = const()[name = tensor("op_270_promoted"), val = tensor(0x1p+0)]; + tensor var_271 = add(x = var_269, y = var_270_promoted)[name = tensor("op_271")]; + tensor var_272_promoted = const()[name = tensor("op_272_promoted"), val = tensor(0x1.8p+1)]; + tensor var_273 = sub(x = var_271, y = var_272_promoted)[name = tensor("op_273")]; + tensor var_53_promoted_2 = const()[name = tensor("op_53_promoted_2"), val = tensor(0x1p+1)]; + tensor floor_div_2 = floor_div(x = var_273, y = var_53_promoted_2)[name = tensor("floor_div_2")]; + tensor var_275_promoted = const()[name = tensor("op_275_promoted"), val = tensor(0x1p+0)]; + tensor current_lengths = add(x = floor_div_2, y = var_275_promoted)[name = tensor("current_lengths")]; + tensor lengths_dtype_0 = const()[name = tensor("lengths_dtype_0"), val = tensor("int32")]; + tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3]])]; + tensor var_284_axes_0 = const()[name = tensor("op_284_axes_0"), val = tensor([1])]; + tensor lengths = cast(dtype = lengths_dtype_0, x = current_lengths)[name = tensor("cast_5")]; + tensor var_284 = expand_dims(axes = var_284_axes_0, x = lengths)[name = tensor("op_284")]; + tensor time_mask = less(x = expand_dims_3, y = var_284)[name = tensor("time_mask")]; + tensor var_286_axes_0 = const()[name = tensor("op_286_axes_0"), val = tensor([-1])]; + tensor var_286 = expand_dims(axes = var_286_axes_0, x = time_mask)[name = tensor("op_286")]; + tensor var_288_reps_0 = const()[name = tensor("op_288_reps_0"), val = tensor([1, 1, 17])]; + tensor var_288 = tile(reps = var_288_reps_0, x = var_286)[name = tensor("op_288")]; + tensor mask_7_dtype_0 = const()[name = tensor("mask_7_dtype_0"), val = tensor("fp32")]; + tensor var_294_axes_0 = const()[name = tensor("op_294_axes_0"), val = tensor([1])]; + tensor mask_7 = cast(dtype = mask_7_dtype_0, x = var_288)[name = tensor("cast_4")]; + tensor var_294 = expand_dims(axes = var_294_axes_0, x = mask_7)[name = tensor("op_294")]; + tensor expanded_mask_13_reps_0 = const()[name = tensor("expanded_mask_13_reps_0"), val = tensor([1, 256, 1, 1])]; + tensor expanded_mask_13 = tile(reps = expanded_mask_13_reps_0, x = var_294)[name = tensor("expanded_mask_13")]; + tensor input_19 = mul(x = tensor_13, y = expanded_mask_13)[name = tensor("input_19")]; + tensor tensor_15_pad_type_0 = const()[name = tensor("tensor_15_pad_type_0"), val = tensor("valid")]; + tensor tensor_15_strides_0 = const()[name = tensor("tensor_15_strides_0"), val = tensor([1, 1])]; + tensor tensor_15_pad_0 = const()[name = tensor("tensor_15_pad_0"), val = tensor([0, 0, 0, 0])]; + tensor tensor_15_dilations_0 = const()[name = tensor("tensor_15_dilations_0"), val = tensor([1, 1])]; + tensor tensor_15_groups_0 = const()[name = tensor("tensor_15_groups_0"), val = tensor(1)]; + tensor tensor_15 = conv(bias = module_pre_encode_conv_6_bias, dilations = tensor_15_dilations_0, groups = tensor_15_groups_0, pad = tensor_15_pad_0, pad_type = tensor_15_pad_type_0, strides = tensor_15_strides_0, weight = module_pre_encode_conv_6_weight_quantized, x = input_19)[name = tensor("tensor_15")]; + tensor input_21 = mul(x = tensor_15, y = expanded_mask_13)[name = tensor("input_21")]; + tensor tensor_workaround = relu(x = input_21)[name = tensor("tensor_workaround")]; + tensor x_3 = mul(x = tensor_workaround, y = expanded_mask_13)[name = tensor("x_3")]; + tensor var_328_perm_0 = const()[name = tensor("op_328_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_329 = const()[name = tensor("op_329"), val = tensor([1, 4, -1])]; + tensor var_328 = transpose(perm = var_328_perm_0, x = x_3)[name = tensor("transpose_366")]; + tensor input_23 = reshape(shape = var_329, x = var_328)[name = tensor("input_23")]; + tensor audio_signal_1 = linear(bias = module_pre_encode_out_bias, weight = module_pre_encode_out_weight_quantized, x = input_23)[name = tensor("linear_0")]; + tensor var_339_begin_0 = const()[name = tensor("op_339_begin_0"), val = tensor([0, 2, 0])]; + tensor var_339_end_0 = const()[name = tensor("op_339_end_0"), val = tensor([1, 4, 1024])]; + tensor var_339_end_mask_0 = const()[name = tensor("op_339_end_mask_0"), val = tensor([true, true, true])]; + tensor var_339 = slice_by_index(begin = var_339_begin_0, end = var_339_end_0, end_mask = var_339_end_mask_0, x = audio_signal_1)[name = tensor("op_339")]; + tensor var_341 = const()[name = tensor("op_341"), val = tensor(2)]; + tensor var_342 = sub(x = lengths, y = var_341)[name = tensor("op_342")]; + tensor const_61 = const()[name = tensor("const_61"), val = tensor(0x1.fffffep+127)]; + tensor var_342_promoted_dtype_0 = const()[name = tensor("op_342_promoted_dtype_0"), val = tensor("fp32")]; + tensor var_59_promoted = const()[name = tensor("op_59_promoted"), val = tensor(0x0p+0)]; + tensor var_342_promoted = cast(dtype = var_342_promoted_dtype_0, x = var_342)[name = tensor("cast_3")]; + tensor clip_0 = clip(alpha = var_59_promoted, beta = const_61, x = var_342_promoted)[name = tensor("clip_0")]; + tensor max_audio_length_1 = const()[name = tensor("max_audio_length_1"), val = tensor([2])]; + tensor var_358_promoted = const()[name = tensor("op_358_promoted"), val = tensor(0x1.18p+6)]; + tensor padding_length = add(x = clip_0, y = var_358_promoted)[name = tensor("padding_length")]; + tensor const_63 = const()[name = tensor("const_63"), val = tensor(-1)]; + tensor var_360 = mul(x = cache_len, y = const_63)[name = tensor("op_360")]; + tensor var_361 = const()[name = tensor("op_361"), val = tensor(70)]; + tensor offset = add(x = var_360, y = var_361)[name = tensor("offset")]; + tensor var_401_axes_0 = const()[name = tensor("op_401_axes_0"), val = tensor([-1])]; + tensor var_401 = expand_dims(axes = var_401_axes_0, x = padding_length)[name = tensor("op_401")]; + tensor var_400_promoted = const()[name = tensor("op_400_promoted"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587312192)))]; + tensor pad_mask_1 = less(x = var_400_promoted, y = var_401)[name = tensor("pad_mask_1")]; + tensor expand_dims_5 = const()[name = tensor("expand_dims_5"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71]])]; + tensor var_407_axes_0 = const()[name = tensor("op_407_axes_0"), val = tensor([-1])]; + tensor var_407 = expand_dims(axes = var_407_axes_0, x = offset)[name = tensor("op_407")]; + tensor pad_mask_off = greater_equal(x = expand_dims_5, y = var_407)[name = tensor("pad_mask_off")]; + tensor pad_mask_3 = logical_and(x = pad_mask_off, y = pad_mask_1)[name = tensor("pad_mask_3")]; + tensor var_410_axes_0 = const()[name = tensor("op_410_axes_0"), val = tensor([1])]; + tensor var_410 = expand_dims(axes = var_410_axes_0, x = pad_mask_3)[name = tensor("op_410")]; + tensor var_411 = const()[name = tensor("op_411"), val = tensor([1, 72, 1])]; + tensor pad_mask_for_att_mask_1 = tile(reps = var_411, x = var_410)[name = tensor("pad_mask_for_att_mask_1")]; + tensor var_413_perm_0 = const()[name = tensor("op_413_perm_0"), val = tensor([0, 2, 1])]; + tensor var_413 = transpose(perm = var_413_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_365")]; + tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_413)[name = tensor("pad_mask_for_att_mask")]; + tensor const_71 = const()[name = tensor("const_71"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, false, false], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; + tensor att_mask_9 = logical_and(x = pad_mask_for_att_mask, y = const_71)[name = tensor("att_mask_9")]; + tensor att_mask = logical_not(x = att_mask_9)[name = tensor("att_mask")]; + tensor pad_mask_5 = logical_not(x = pad_mask_3)[name = tensor("pad_mask_5")]; + tensor pad_mask_begin_0 = const()[name = tensor("pad_mask_begin_0"), val = tensor([0, 70])]; + tensor pad_mask_end_0 = const()[name = tensor("pad_mask_end_0"), val = tensor([1, 72])]; + tensor pad_mask_end_mask_0 = const()[name = tensor("pad_mask_end_mask_0"), val = tensor([true, true])]; + tensor pad_mask = slice_by_index(begin = pad_mask_begin_0, end = pad_mask_end_0, end_mask = pad_mask_end_mask_0, x = pad_mask_5)[name = tensor("pad_mask")]; + tensor mask_9_begin_0 = const()[name = tensor("mask_9_begin_0"), val = tensor([0, 70, 0])]; + tensor mask_9_end_0 = const()[name = tensor("mask_9_end_0"), val = tensor([1, 72, 72])]; + tensor mask_9_end_mask_0 = const()[name = tensor("mask_9_end_mask_0"), val = tensor([true, true, true])]; + tensor mask_9 = slice_by_index(begin = mask_9_begin_0, end = mask_9_end_0, end_mask = mask_9_end_mask_0, x = att_mask)[name = tensor("mask_9")]; + tensor cache_1_begin_0 = const()[name = tensor("cache_1_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_1_end_0 = const()[name = tensor("cache_1_end_0"), val = tensor([1, 1, 70, 1024])]; + tensor cache_1_end_mask_0 = const()[name = tensor("cache_1_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_1_squeeze_mask_0 = const()[name = tensor("cache_1_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor value_3 = transpose(perm = value_3_perm_0, x = cache_channel)[name = tensor("transpose_364")]; + tensor cache_1 = slice_by_index(begin = cache_1_begin_0, end = cache_1_end_0, end_mask = cache_1_end_mask_0, squeeze_mask = cache_1_squeeze_mask_0, x = value_3)[name = tensor("cache_1")]; + tensor cache_3_begin_0 = const()[name = tensor("cache_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor cache_3_end_0 = const()[name = tensor("cache_3_end_0"), val = tensor([1, 1, 1024, 8])]; + tensor cache_3_end_mask_0 = const()[name = tensor("cache_3_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_3_squeeze_mask_0 = const()[name = tensor("cache_3_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor value_5 = transpose(perm = value_5_perm_0, x = cache_time)[name = tensor("transpose_363")]; + tensor cache_3 = slice_by_index(begin = cache_3_begin_0, end = cache_3_end_0, end_mask = cache_3_end_mask_0, squeeze_mask = cache_3_squeeze_mask_0, x = value_5)[name = tensor("cache_3")]; + tensor input_27_axes_0 = const()[name = tensor("input_27_axes_0"), val = tensor([-1])]; + tensor input_27 = layer_norm(axes = input_27_axes_0, beta = module_layers_0_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_0_norm_feed_forward1_weight, x = var_339)[name = tensor("input_27")]; + tensor linear_1_bias_0 = const()[name = tensor("linear_1_bias_0"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587312576)))]; + tensor input_29 = linear(bias = linear_1_bias_0, weight = module_layers_0_feed_forward1_linear1_weight_quantized, x = input_27)[name = tensor("linear_1")]; + tensor input_31 = silu(x = input_29)[name = tensor("input_31")]; + tensor linear_2_bias_0 = const()[name = tensor("linear_2_bias_0"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587329024)))]; + tensor input_35 = linear(bias = linear_2_bias_0, weight = module_layers_0_feed_forward1_linear2_weight_quantized, x = input_31)[name = tensor("linear_2")]; + tensor var_450 = const()[name = tensor("op_450"), val = tensor(0x1p-1)]; + tensor var_451 = mul(x = input_35, y = var_450)[name = tensor("op_451")]; + tensor input_37 = add(x = var_339, y = var_451)[name = tensor("input_37")]; + tensor key_1_axes_0 = const()[name = tensor("key_1_axes_0"), val = tensor([-1])]; + tensor key_1 = layer_norm(axes = key_1_axes_0, beta = module_layers_0_norm_self_att_bias, epsilon = var_38, gamma = module_layers_0_norm_self_att_weight, x = input_37)[name = tensor("key_1")]; + tensor input_39_interleave_0 = const()[name = tensor("input_39_interleave_0"), val = tensor(false)]; + tensor input_39 = concat(axis = var_65, interleave = input_39_interleave_0, values = (cache_1, key_1))[name = tensor("input_39")]; + tensor var_473_begin_0 = const()[name = tensor("op_473_begin_0"), val = tensor([0, 2, 0])]; + tensor var_473_end_0 = const()[name = tensor("op_473_end_0"), val = tensor([1, 70, 1024])]; + tensor var_473_end_mask_0 = const()[name = tensor("op_473_end_mask_0"), val = tensor([true, true, true])]; + tensor var_473 = slice_by_index(begin = var_473_begin_0, end = var_473_end_0, end_mask = var_473_end_mask_0, x = cache_1)[name = tensor("op_473")]; + tensor var_479_interleave_0 = const()[name = tensor("op_479_interleave_0"), val = tensor(false)]; + tensor var_479 = concat(axis = var_65, interleave = var_479_interleave_0, values = (var_473, key_1))[name = tensor("op_479")]; + tensor var_482 = linear(bias = linear_2_bias_0, weight = module_layers_0_self_attn_linear_q_weight_quantized, x = key_1)[name = tensor("linear_3")]; + tensor var_483 = const()[name = tensor("op_483"), val = tensor([1, -1, 8, 128])]; + tensor q_1 = reshape(shape = var_483, x = var_482)[name = tensor("q_1")]; + tensor var_486 = linear(bias = linear_2_bias_0, weight = module_layers_0_self_attn_linear_k_weight_quantized, x = input_39)[name = tensor("linear_4")]; + tensor var_487 = const()[name = tensor("op_487"), val = tensor([1, -1, 8, 128])]; + tensor k_1 = reshape(shape = var_487, x = var_486)[name = tensor("k_1")]; + tensor var_490 = linear(bias = linear_2_bias_0, weight = module_layers_0_self_attn_linear_v_weight_quantized, x = input_39)[name = tensor("linear_5")]; + tensor var_491 = const()[name = tensor("op_491"), val = tensor([1, -1, 8, 128])]; + tensor v_1 = reshape(shape = var_491, x = var_490)[name = tensor("v_1")]; + tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_503 = add(x = q_1, y = module_layers_0_self_attn_pos_bias_u)[name = tensor("op_503")]; + tensor var_505 = add(x = q_1, y = module_layers_0_self_attn_pos_bias_v)[name = tensor("op_505")]; + tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_507_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_507_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587333184))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587479936))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587479680)))]; + tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; + tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_1 = transpose(perm = q_with_bias_v_1_perm_0, x = var_505)[name = tensor("transpose_362")]; + tensor x_7 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1, y = op_507_quantized)[name = tensor("x_7")]; + tensor const_79 = const()[name = tensor("const_79"), val = tensor(0x0p+0)]; + tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; + tensor x_9 = pad(constant_val = const_79, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7)[name = tensor("x_9")]; + tensor var_515 = const()[name = tensor("op_515"), val = tensor([1, 8, -1, 2])]; + tensor x_11 = reshape(shape = var_515, x = x_9)[name = tensor("x_11")]; + tensor var_519_begin_0 = const()[name = tensor("op_519_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_519_end_0 = const()[name = tensor("op_519_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_519_end_mask_0 = const()[name = tensor("op_519_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_519 = slice_by_index(begin = var_519_begin_0, end = var_519_end_0, end_mask = var_519_end_mask_0, x = x_11)[name = tensor("op_519")]; + tensor var_520 = const()[name = tensor("op_520"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_1 = reshape(shape = var_520, x = var_519)[name = tensor("matrix_bd_1")]; + tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; + tensor transpose_96_perm_0 = const()[name = tensor("transpose_96_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_97_perm_0 = const()[name = tensor("transpose_97_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_97 = transpose(perm = transpose_97_perm_0, x = k_1)[name = tensor("transpose_360")]; + tensor transpose_96 = transpose(perm = transpose_96_perm_0, x = var_503)[name = tensor("transpose_361")]; + tensor matrix_ac_1 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_96, y = transpose_97)[name = tensor("matrix_ac_1")]; + tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_3 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1)[name = tensor("matrix_bd_3")]; + tensor var_529 = add(x = matrix_ac_1, y = matrix_bd_3)[name = tensor("op_529")]; + tensor _inversed_scores_1_y_0 = const()[name = tensor("_inversed_scores_1_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_1 = mul(x = var_529, y = _inversed_scores_1_y_0)[name = tensor("_inversed_scores_1")]; + tensor mask_11_axes_0 = const()[name = tensor("mask_11_axes_0"), val = tensor([1])]; + tensor mask_11 = expand_dims(axes = mask_11_axes_0, x = mask_9)[name = tensor("mask_11")]; + tensor scores_3 = select(a = var_41, b = _inversed_scores_1, cond = mask_11)[name = tensor("scores_3")]; + tensor var_535 = softmax(axis = var_56, x = scores_3)[name = tensor("op_535")]; + tensor input_41 = select(a = var_40, b = var_535, cond = mask_11)[name = tensor("input_41")]; + tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; + tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; + tensor value_9 = transpose(perm = value_9_perm_0, x = v_1)[name = tensor("transpose_359")]; + tensor x_13 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_41, y = value_9)[name = tensor("x_13")]; + tensor var_539_perm_0 = const()[name = tensor("op_539_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_540 = const()[name = tensor("op_540"), val = tensor([1, -1, 1024])]; + tensor var_539 = transpose(perm = var_539_perm_0, x = x_13)[name = tensor("transpose_358")]; + tensor input_43 = reshape(shape = var_540, x = var_539)[name = tensor("input_43")]; + tensor input_45 = linear(bias = linear_2_bias_0, weight = module_layers_0_self_attn_linear_out_weight_quantized, x = input_43)[name = tensor("linear_7")]; + tensor input_47 = add(x = input_37, y = input_45)[name = tensor("input_47")]; + tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; + tensor x_17 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias, epsilon = var_38, gamma = module_layers_0_norm_conv_weight, x = input_47)[name = tensor("x_17")]; + tensor input_49_perm_0 = const()[name = tensor("input_49_perm_0"), val = tensor([0, 2, 1])]; + tensor input_51_pad_type_0 = const()[name = tensor("input_51_pad_type_0"), val = tensor("valid")]; + tensor input_51_strides_0 = const()[name = tensor("input_51_strides_0"), val = tensor([1])]; + tensor input_51_pad_0 = const()[name = tensor("input_51_pad_0"), val = tensor([0, 0])]; + tensor input_51_dilations_0 = const()[name = tensor("input_51_dilations_0"), val = tensor([1])]; + tensor input_51_groups_0 = const()[name = tensor("input_51_groups_0"), val = tensor(1)]; + tensor input_49 = transpose(perm = input_49_perm_0, x = x_17)[name = tensor("transpose_357")]; + tensor input_51 = conv(dilations = input_51_dilations_0, groups = input_51_groups_0, pad = input_51_pad_0, pad_type = input_51_pad_type_0, strides = input_51_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_quantized, x = input_49)[name = tensor("input_51")]; + tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; + tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; + tensor x_19_split_0, tensor x_19_split_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_51)[name = tensor("x_19_split")]; + tensor x_19_split_1_sigmoid = sigmoid(x = x_19_split_1)[name = tensor("x_19_split_1_sigmoid")]; + tensor x_19 = mul(x = x_19_split_0, y = x_19_split_1_sigmoid)[name = tensor("x_19")]; + tensor var_565_axes_0 = const()[name = tensor("op_565_axes_0"), val = tensor([1])]; + tensor var_565 = expand_dims(axes = var_565_axes_0, x = pad_mask)[name = tensor("op_565")]; + tensor input_53 = select(a = var_40, b = x_19, cond = var_565)[name = tensor("input_53")]; + tensor new_x_3_interleave_0 = const()[name = tensor("new_x_3_interleave_0"), val = tensor(false)]; + tensor new_x_3 = concat(axis = var_56, interleave = new_x_3_interleave_0, values = (cache_3, input_53))[name = tensor("new_x_3")]; + tensor var_578_begin_0 = const()[name = tensor("op_578_begin_0"), val = tensor([0, 0, 2])]; + tensor var_578_end_0 = const()[name = tensor("op_578_end_0"), val = tensor([1, 1024, 10])]; + tensor var_578_end_mask_0 = const()[name = tensor("op_578_end_mask_0"), val = tensor([true, true, true])]; + tensor var_578 = slice_by_index(begin = var_578_begin_0, end = var_578_end_0, end_mask = var_578_end_mask_0, x = new_x_3)[name = tensor("op_578")]; + tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; + tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1024)]; + tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; + tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; + tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; + tensor x_21 = conv(dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_depthwise_conv_weight_quantized, x = new_x_3)[name = tensor("x_21")]; + tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; + tensor x_23_axes_0 = const()[name = tensor("x_23_axes_0"), val = tensor([-1])]; + tensor input_55 = transpose(perm = input_55_perm_0, x = x_21)[name = tensor("transpose_356")]; + tensor x_23 = layer_norm(axes = x_23_axes_0, beta = module_layers_0_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_0_conv_batch_norm_weight, x = input_55)[name = tensor("x_23")]; + tensor input_57_perm_0 = const()[name = tensor("input_57_perm_0"), val = tensor([0, 2, 1])]; + tensor input_57 = transpose(perm = input_57_perm_0, x = x_23)[name = tensor("transpose_355")]; + tensor input_59 = silu(x = input_57)[name = tensor("input_59")]; + tensor x_25_pad_type_0 = const()[name = tensor("x_25_pad_type_0"), val = tensor("valid")]; + tensor x_25_strides_0 = const()[name = tensor("x_25_strides_0"), val = tensor([1])]; + tensor x_25_pad_0 = const()[name = tensor("x_25_pad_0"), val = tensor([0, 0])]; + tensor x_25_dilations_0 = const()[name = tensor("x_25_dilations_0"), val = tensor([1])]; + tensor x_25_groups_0 = const()[name = tensor("x_25_groups_0"), val = tensor(1)]; + tensor x_25 = conv(dilations = x_25_dilations_0, groups = x_25_groups_0, pad = x_25_pad_0, pad_type = x_25_pad_type_0, strides = x_25_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_quantized, x = input_59)[name = tensor("x_25")]; + tensor input_61_perm_0 = const()[name = tensor("input_61_perm_0"), val = tensor([0, 2, 1])]; + tensor input_61 = transpose(perm = input_61_perm_0, x = x_25)[name = tensor("transpose_354")]; + tensor input_63 = add(x = input_47, y = input_61)[name = tensor("input_63")]; + tensor input_65_axes_0 = const()[name = tensor("input_65_axes_0"), val = tensor([-1])]; + tensor input_65 = layer_norm(axes = input_65_axes_0, beta = module_layers_0_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_0_norm_feed_forward2_weight, x = input_63)[name = tensor("input_65")]; + tensor input_67 = linear(bias = linear_1_bias_0, weight = module_layers_0_feed_forward2_linear1_weight_quantized, x = input_65)[name = tensor("linear_8")]; + tensor input_69 = silu(x = input_67)[name = tensor("input_69")]; + tensor input_73 = linear(bias = linear_2_bias_0, weight = module_layers_0_feed_forward2_linear2_weight_quantized, x = input_69)[name = tensor("linear_9")]; + tensor var_619 = const()[name = tensor("op_619"), val = tensor(0x1p-1)]; + tensor var_620 = mul(x = input_73, y = var_619)[name = tensor("op_620")]; + tensor input_75 = add(x = input_63, y = var_620)[name = tensor("input_75")]; + tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; + tensor input_77 = layer_norm(axes = input_77_axes_0, beta = module_layers_0_norm_out_bias, epsilon = var_38, gamma = module_layers_0_norm_out_weight, x = input_75)[name = tensor("input_77")]; + tensor cache_5_begin_0 = const()[name = tensor("cache_5_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_5_end_0 = const()[name = tensor("cache_5_end_0"), val = tensor([2, 1, 70, 1024])]; + tensor cache_5_end_mask_0 = const()[name = tensor("cache_5_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_5_squeeze_mask_0 = const()[name = tensor("cache_5_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_5 = slice_by_index(begin = cache_5_begin_0, end = cache_5_end_0, end_mask = cache_5_end_mask_0, squeeze_mask = cache_5_squeeze_mask_0, x = value_3)[name = tensor("cache_5")]; + tensor cache_7_begin_0 = const()[name = tensor("cache_7_begin_0"), val = tensor([1, 0, 0, 0])]; + tensor cache_7_end_0 = const()[name = tensor("cache_7_end_0"), val = tensor([2, 1, 1024, 8])]; + tensor cache_7_end_mask_0 = const()[name = tensor("cache_7_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_7_squeeze_mask_0 = const()[name = tensor("cache_7_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_7 = slice_by_index(begin = cache_7_begin_0, end = cache_7_end_0, end_mask = cache_7_end_mask_0, squeeze_mask = cache_7_squeeze_mask_0, x = value_5)[name = tensor("cache_7")]; + tensor input_79_axes_0 = const()[name = tensor("input_79_axes_0"), val = tensor([-1])]; + tensor input_79 = layer_norm(axes = input_79_axes_0, beta = module_layers_1_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_1_norm_feed_forward1_weight, x = input_77)[name = tensor("input_79")]; + tensor input_81 = linear(bias = linear_1_bias_0, weight = module_layers_1_feed_forward1_linear1_weight_quantized, x = input_79)[name = tensor("linear_10")]; + tensor input_83 = silu(x = input_81)[name = tensor("input_83")]; + tensor input_87 = linear(bias = linear_2_bias_0, weight = module_layers_1_feed_forward1_linear2_weight_quantized, x = input_83)[name = tensor("linear_11")]; + tensor var_654 = const()[name = tensor("op_654"), val = tensor(0x1p-1)]; + tensor var_655 = mul(x = input_87, y = var_654)[name = tensor("op_655")]; + tensor input_89 = add(x = input_77, y = var_655)[name = tensor("input_89")]; + tensor key_3_axes_0 = const()[name = tensor("key_3_axes_0"), val = tensor([-1])]; + tensor key_3 = layer_norm(axes = key_3_axes_0, beta = module_layers_1_norm_self_att_bias, epsilon = var_38, gamma = module_layers_1_norm_self_att_weight, x = input_89)[name = tensor("key_3")]; + tensor input_91_interleave_0 = const()[name = tensor("input_91_interleave_0"), val = tensor(false)]; + tensor input_91 = concat(axis = var_65, interleave = input_91_interleave_0, values = (cache_5, key_3))[name = tensor("input_91")]; + tensor var_677_begin_0 = const()[name = tensor("op_677_begin_0"), val = tensor([0, 2, 0])]; + tensor var_677_end_0 = const()[name = tensor("op_677_end_0"), val = tensor([1, 70, 1024])]; + tensor var_677_end_mask_0 = const()[name = tensor("op_677_end_mask_0"), val = tensor([true, true, true])]; + tensor var_677 = slice_by_index(begin = var_677_begin_0, end = var_677_end_0, end_mask = var_677_end_mask_0, x = cache_5)[name = tensor("op_677")]; + tensor var_683_interleave_0 = const()[name = tensor("op_683_interleave_0"), val = tensor(false)]; + tensor var_683 = concat(axis = var_65, interleave = var_683_interleave_0, values = (var_677, key_3))[name = tensor("op_683")]; + tensor var_686 = linear(bias = linear_2_bias_0, weight = module_layers_1_self_attn_linear_q_weight_quantized, x = key_3)[name = tensor("linear_12")]; + tensor var_687 = const()[name = tensor("op_687"), val = tensor([1, -1, 8, 128])]; + tensor q_7 = reshape(shape = var_687, x = var_686)[name = tensor("q_7")]; + tensor var_690 = linear(bias = linear_2_bias_0, weight = module_layers_1_self_attn_linear_k_weight_quantized, x = input_91)[name = tensor("linear_13")]; + tensor var_691 = const()[name = tensor("op_691"), val = tensor([1, -1, 8, 128])]; + tensor k_5 = reshape(shape = var_691, x = var_690)[name = tensor("k_5")]; + tensor var_694 = linear(bias = linear_2_bias_0, weight = module_layers_1_self_attn_linear_v_weight_quantized, x = input_91)[name = tensor("linear_14")]; + tensor var_695 = const()[name = tensor("op_695"), val = tensor([1, -1, 8, 128])]; + tensor v_3 = reshape(shape = var_695, x = var_694)[name = tensor("v_3")]; + tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_707 = add(x = q_7, y = module_layers_1_self_attn_pos_bias_u)[name = tensor("op_707")]; + tensor var_709 = add(x = q_7, y = module_layers_1_self_attn_pos_bias_v)[name = tensor("op_709")]; + tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_711_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_711_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587480576))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587627328))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587627072)))]; + tensor x_33_transpose_x_0 = const()[name = tensor("x_33_transpose_x_0"), val = tensor(false)]; + tensor x_33_transpose_y_0 = const()[name = tensor("x_33_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_3 = transpose(perm = q_with_bias_v_3_perm_0, x = var_709)[name = tensor("transpose_353")]; + tensor x_33 = matmul(transpose_x = x_33_transpose_x_0, transpose_y = x_33_transpose_y_0, x = q_with_bias_v_3, y = op_711_quantized)[name = tensor("x_33")]; + tensor const_92 = const()[name = tensor("const_92"), val = tensor(0x0p+0)]; + tensor x_35_pad_0 = const()[name = tensor("x_35_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_35_mode_0 = const()[name = tensor("x_35_mode_0"), val = tensor("constant")]; + tensor x_35 = pad(constant_val = const_92, mode = x_35_mode_0, pad = x_35_pad_0, x = x_33)[name = tensor("x_35")]; + tensor var_719 = const()[name = tensor("op_719"), val = tensor([1, 8, -1, 2])]; + tensor x_37 = reshape(shape = var_719, x = x_35)[name = tensor("x_37")]; + tensor var_723_begin_0 = const()[name = tensor("op_723_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_723_end_0 = const()[name = tensor("op_723_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_723_end_mask_0 = const()[name = tensor("op_723_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_723 = slice_by_index(begin = var_723_begin_0, end = var_723_end_0, end_mask = var_723_end_mask_0, x = x_37)[name = tensor("op_723")]; + tensor var_724 = const()[name = tensor("op_724"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_5 = reshape(shape = var_724, x = var_723)[name = tensor("matrix_bd_5")]; + tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; + tensor transpose_98_perm_0 = const()[name = tensor("transpose_98_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_99_perm_0 = const()[name = tensor("transpose_99_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_99 = transpose(perm = transpose_99_perm_0, x = k_5)[name = tensor("transpose_351")]; + tensor transpose_98 = transpose(perm = transpose_98_perm_0, x = var_707)[name = tensor("transpose_352")]; + tensor matrix_ac_3 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_98, y = transpose_99)[name = tensor("matrix_ac_3")]; + tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_7 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5)[name = tensor("matrix_bd_7")]; + tensor var_733 = add(x = matrix_ac_3, y = matrix_bd_7)[name = tensor("op_733")]; + tensor _inversed_scores_5_y_0 = const()[name = tensor("_inversed_scores_5_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_5 = mul(x = var_733, y = _inversed_scores_5_y_0)[name = tensor("_inversed_scores_5")]; + tensor scores_7 = select(a = var_41, b = _inversed_scores_5, cond = mask_11)[name = tensor("scores_7")]; + tensor var_739 = softmax(axis = var_56, x = scores_7)[name = tensor("op_739")]; + tensor input_93 = select(a = var_40, b = var_739, cond = mask_11)[name = tensor("input_93")]; + tensor x_39_transpose_x_0 = const()[name = tensor("x_39_transpose_x_0"), val = tensor(false)]; + tensor x_39_transpose_y_0 = const()[name = tensor("x_39_transpose_y_0"), val = tensor(false)]; + tensor value_11 = transpose(perm = value_11_perm_0, x = v_3)[name = tensor("transpose_350")]; + tensor x_39 = matmul(transpose_x = x_39_transpose_x_0, transpose_y = x_39_transpose_y_0, x = input_93, y = value_11)[name = tensor("x_39")]; + tensor var_743_perm_0 = const()[name = tensor("op_743_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_744 = const()[name = tensor("op_744"), val = tensor([1, -1, 1024])]; + tensor var_743 = transpose(perm = var_743_perm_0, x = x_39)[name = tensor("transpose_349")]; + tensor input_95 = reshape(shape = var_744, x = var_743)[name = tensor("input_95")]; + tensor input_97 = linear(bias = linear_2_bias_0, weight = module_layers_1_self_attn_linear_out_weight_quantized, x = input_95)[name = tensor("linear_16")]; + tensor input_99 = add(x = input_89, y = input_97)[name = tensor("input_99")]; + tensor x_43_axes_0 = const()[name = tensor("x_43_axes_0"), val = tensor([-1])]; + tensor x_43 = layer_norm(axes = x_43_axes_0, beta = module_layers_1_norm_conv_bias, epsilon = var_38, gamma = module_layers_1_norm_conv_weight, x = input_99)[name = tensor("x_43")]; + tensor input_101_perm_0 = const()[name = tensor("input_101_perm_0"), val = tensor([0, 2, 1])]; + tensor input_103_pad_type_0 = const()[name = tensor("input_103_pad_type_0"), val = tensor("valid")]; + tensor input_103_strides_0 = const()[name = tensor("input_103_strides_0"), val = tensor([1])]; + tensor input_103_pad_0 = const()[name = tensor("input_103_pad_0"), val = tensor([0, 0])]; + tensor input_103_dilations_0 = const()[name = tensor("input_103_dilations_0"), val = tensor([1])]; + tensor input_103_groups_0 = const()[name = tensor("input_103_groups_0"), val = tensor(1)]; + tensor input_101 = transpose(perm = input_101_perm_0, x = x_43)[name = tensor("transpose_348")]; + tensor input_103 = conv(dilations = input_103_dilations_0, groups = input_103_groups_0, pad = input_103_pad_0, pad_type = input_103_pad_type_0, strides = input_103_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_quantized, x = input_101)[name = tensor("input_103")]; + tensor x_45_split_num_splits_0 = const()[name = tensor("x_45_split_num_splits_0"), val = tensor(2)]; + tensor x_45_split_axis_0 = const()[name = tensor("x_45_split_axis_0"), val = tensor(1)]; + tensor x_45_split_0, tensor x_45_split_1 = split(axis = x_45_split_axis_0, num_splits = x_45_split_num_splits_0, x = input_103)[name = tensor("x_45_split")]; + tensor x_45_split_1_sigmoid = sigmoid(x = x_45_split_1)[name = tensor("x_45_split_1_sigmoid")]; + tensor x_45 = mul(x = x_45_split_0, y = x_45_split_1_sigmoid)[name = tensor("x_45")]; + tensor input_105 = select(a = var_40, b = x_45, cond = var_565)[name = tensor("input_105")]; + tensor new_x_7_interleave_0 = const()[name = tensor("new_x_7_interleave_0"), val = tensor(false)]; + tensor new_x_7 = concat(axis = var_56, interleave = new_x_7_interleave_0, values = (cache_7, input_105))[name = tensor("new_x_7")]; + tensor var_782_begin_0 = const()[name = tensor("op_782_begin_0"), val = tensor([0, 0, 2])]; + tensor var_782_end_0 = const()[name = tensor("op_782_end_0"), val = tensor([1, 1024, 10])]; + tensor var_782_end_mask_0 = const()[name = tensor("op_782_end_mask_0"), val = tensor([true, true, true])]; + tensor var_782 = slice_by_index(begin = var_782_begin_0, end = var_782_end_0, end_mask = var_782_end_mask_0, x = new_x_7)[name = tensor("op_782")]; + tensor x_47_pad_type_0 = const()[name = tensor("x_47_pad_type_0"), val = tensor("valid")]; + tensor x_47_groups_0 = const()[name = tensor("x_47_groups_0"), val = tensor(1024)]; + tensor x_47_strides_0 = const()[name = tensor("x_47_strides_0"), val = tensor([1])]; + tensor x_47_pad_0 = const()[name = tensor("x_47_pad_0"), val = tensor([0, 0])]; + tensor x_47_dilations_0 = const()[name = tensor("x_47_dilations_0"), val = tensor([1])]; + tensor x_47 = conv(dilations = x_47_dilations_0, groups = x_47_groups_0, pad = x_47_pad_0, pad_type = x_47_pad_type_0, strides = x_47_strides_0, weight = module_layers_1_conv_depthwise_conv_weight_quantized, x = new_x_7)[name = tensor("x_47")]; + tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; + tensor x_49_axes_0 = const()[name = tensor("x_49_axes_0"), val = tensor([-1])]; + tensor input_107 = transpose(perm = input_107_perm_0, x = x_47)[name = tensor("transpose_347")]; + tensor x_49 = layer_norm(axes = x_49_axes_0, beta = module_layers_1_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_1_conv_batch_norm_weight, x = input_107)[name = tensor("x_49")]; + tensor input_109_perm_0 = const()[name = tensor("input_109_perm_0"), val = tensor([0, 2, 1])]; + tensor input_109 = transpose(perm = input_109_perm_0, x = x_49)[name = tensor("transpose_346")]; + tensor input_111 = silu(x = input_109)[name = tensor("input_111")]; + tensor x_51_pad_type_0 = const()[name = tensor("x_51_pad_type_0"), val = tensor("valid")]; + tensor x_51_strides_0 = const()[name = tensor("x_51_strides_0"), val = tensor([1])]; + tensor x_51_pad_0 = const()[name = tensor("x_51_pad_0"), val = tensor([0, 0])]; + tensor x_51_dilations_0 = const()[name = tensor("x_51_dilations_0"), val = tensor([1])]; + tensor x_51_groups_0 = const()[name = tensor("x_51_groups_0"), val = tensor(1)]; + tensor x_51 = conv(dilations = x_51_dilations_0, groups = x_51_groups_0, pad = x_51_pad_0, pad_type = x_51_pad_type_0, strides = x_51_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_quantized, x = input_111)[name = tensor("x_51")]; + tensor input_113_perm_0 = const()[name = tensor("input_113_perm_0"), val = tensor([0, 2, 1])]; + tensor input_113 = transpose(perm = input_113_perm_0, x = x_51)[name = tensor("transpose_345")]; + tensor input_115 = add(x = input_99, y = input_113)[name = tensor("input_115")]; + tensor input_117_axes_0 = const()[name = tensor("input_117_axes_0"), val = tensor([-1])]; + tensor input_117 = layer_norm(axes = input_117_axes_0, beta = module_layers_1_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_1_norm_feed_forward2_weight, x = input_115)[name = tensor("input_117")]; + tensor input_119 = linear(bias = linear_1_bias_0, weight = module_layers_1_feed_forward2_linear1_weight_quantized, x = input_117)[name = tensor("linear_17")]; + tensor input_121 = silu(x = input_119)[name = tensor("input_121")]; + tensor input_125 = linear(bias = linear_2_bias_0, weight = module_layers_1_feed_forward2_linear2_weight_quantized, x = input_121)[name = tensor("linear_18")]; + tensor var_823 = const()[name = tensor("op_823"), val = tensor(0x1p-1)]; + tensor var_824 = mul(x = input_125, y = var_823)[name = tensor("op_824")]; + tensor input_127 = add(x = input_115, y = var_824)[name = tensor("input_127")]; + tensor input_129_axes_0 = const()[name = tensor("input_129_axes_0"), val = tensor([-1])]; + tensor input_129 = layer_norm(axes = input_129_axes_0, beta = module_layers_1_norm_out_bias, epsilon = var_38, gamma = module_layers_1_norm_out_weight, x = input_127)[name = tensor("input_129")]; + tensor cache_9_begin_0 = const()[name = tensor("cache_9_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_9_end_0 = const()[name = tensor("cache_9_end_0"), val = tensor([3, 1, 70, 1024])]; + tensor cache_9_end_mask_0 = const()[name = tensor("cache_9_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_9_squeeze_mask_0 = const()[name = tensor("cache_9_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_9 = slice_by_index(begin = cache_9_begin_0, end = cache_9_end_0, end_mask = cache_9_end_mask_0, squeeze_mask = cache_9_squeeze_mask_0, x = value_3)[name = tensor("cache_9")]; + tensor cache_11_begin_0 = const()[name = tensor("cache_11_begin_0"), val = tensor([2, 0, 0, 0])]; + tensor cache_11_end_0 = const()[name = tensor("cache_11_end_0"), val = tensor([3, 1, 1024, 8])]; + tensor cache_11_end_mask_0 = const()[name = tensor("cache_11_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_11_squeeze_mask_0 = const()[name = tensor("cache_11_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_11 = slice_by_index(begin = cache_11_begin_0, end = cache_11_end_0, end_mask = cache_11_end_mask_0, squeeze_mask = cache_11_squeeze_mask_0, x = value_5)[name = tensor("cache_11")]; + tensor input_131_axes_0 = const()[name = tensor("input_131_axes_0"), val = tensor([-1])]; + tensor input_131 = layer_norm(axes = input_131_axes_0, beta = module_layers_2_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_2_norm_feed_forward1_weight, x = input_129)[name = tensor("input_131")]; + tensor input_133 = linear(bias = linear_1_bias_0, weight = module_layers_2_feed_forward1_linear1_weight_quantized, x = input_131)[name = tensor("linear_19")]; + tensor input_135 = silu(x = input_133)[name = tensor("input_135")]; + tensor input_139 = linear(bias = linear_2_bias_0, weight = module_layers_2_feed_forward1_linear2_weight_quantized, x = input_135)[name = tensor("linear_20")]; + tensor var_858 = const()[name = tensor("op_858"), val = tensor(0x1p-1)]; + tensor var_859 = mul(x = input_139, y = var_858)[name = tensor("op_859")]; + tensor input_141 = add(x = input_129, y = var_859)[name = tensor("input_141")]; + tensor key_5_axes_0 = const()[name = tensor("key_5_axes_0"), val = tensor([-1])]; + tensor key_5 = layer_norm(axes = key_5_axes_0, beta = module_layers_2_norm_self_att_bias, epsilon = var_38, gamma = module_layers_2_norm_self_att_weight, x = input_141)[name = tensor("key_5")]; + tensor input_143_interleave_0 = const()[name = tensor("input_143_interleave_0"), val = tensor(false)]; + tensor input_143 = concat(axis = var_65, interleave = input_143_interleave_0, values = (cache_9, key_5))[name = tensor("input_143")]; + tensor var_881_begin_0 = const()[name = tensor("op_881_begin_0"), val = tensor([0, 2, 0])]; + tensor var_881_end_0 = const()[name = tensor("op_881_end_0"), val = tensor([1, 70, 1024])]; + tensor var_881_end_mask_0 = const()[name = tensor("op_881_end_mask_0"), val = tensor([true, true, true])]; + tensor var_881 = slice_by_index(begin = var_881_begin_0, end = var_881_end_0, end_mask = var_881_end_mask_0, x = cache_9)[name = tensor("op_881")]; + tensor var_887_interleave_0 = const()[name = tensor("op_887_interleave_0"), val = tensor(false)]; + tensor var_887 = concat(axis = var_65, interleave = var_887_interleave_0, values = (var_881, key_5))[name = tensor("op_887")]; + tensor var_890 = linear(bias = linear_2_bias_0, weight = module_layers_2_self_attn_linear_q_weight_quantized, x = key_5)[name = tensor("linear_21")]; + tensor var_891 = const()[name = tensor("op_891"), val = tensor([1, -1, 8, 128])]; + tensor q_13 = reshape(shape = var_891, x = var_890)[name = tensor("q_13")]; + tensor var_894 = linear(bias = linear_2_bias_0, weight = module_layers_2_self_attn_linear_k_weight_quantized, x = input_143)[name = tensor("linear_22")]; + tensor var_895 = const()[name = tensor("op_895"), val = tensor([1, -1, 8, 128])]; + tensor k_9 = reshape(shape = var_895, x = var_894)[name = tensor("k_9")]; + tensor var_898 = linear(bias = linear_2_bias_0, weight = module_layers_2_self_attn_linear_v_weight_quantized, x = input_143)[name = tensor("linear_23")]; + tensor var_899 = const()[name = tensor("op_899"), val = tensor([1, -1, 8, 128])]; + tensor v_5 = reshape(shape = var_899, x = var_898)[name = tensor("v_5")]; + tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_911 = add(x = q_13, y = module_layers_2_self_attn_pos_bias_u)[name = tensor("op_911")]; + tensor var_913 = add(x = q_13, y = module_layers_2_self_attn_pos_bias_v)[name = tensor("op_913")]; + tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_915_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_915_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587627968))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587774720))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587774464)))]; + tensor x_59_transpose_x_0 = const()[name = tensor("x_59_transpose_x_0"), val = tensor(false)]; + tensor x_59_transpose_y_0 = const()[name = tensor("x_59_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_5 = transpose(perm = q_with_bias_v_5_perm_0, x = var_913)[name = tensor("transpose_344")]; + tensor x_59 = matmul(transpose_x = x_59_transpose_x_0, transpose_y = x_59_transpose_y_0, x = q_with_bias_v_5, y = op_915_quantized)[name = tensor("x_59")]; + tensor const_105 = const()[name = tensor("const_105"), val = tensor(0x0p+0)]; + tensor x_61_pad_0 = const()[name = tensor("x_61_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_61_mode_0 = const()[name = tensor("x_61_mode_0"), val = tensor("constant")]; + tensor x_61 = pad(constant_val = const_105, mode = x_61_mode_0, pad = x_61_pad_0, x = x_59)[name = tensor("x_61")]; + tensor var_923 = const()[name = tensor("op_923"), val = tensor([1, 8, -1, 2])]; + tensor x_63 = reshape(shape = var_923, x = x_61)[name = tensor("x_63")]; + tensor var_927_begin_0 = const()[name = tensor("op_927_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_927_end_0 = const()[name = tensor("op_927_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_927_end_mask_0 = const()[name = tensor("op_927_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_927 = slice_by_index(begin = var_927_begin_0, end = var_927_end_0, end_mask = var_927_end_mask_0, x = x_63)[name = tensor("op_927")]; + tensor var_928 = const()[name = tensor("op_928"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_9 = reshape(shape = var_928, x = var_927)[name = tensor("matrix_bd_9")]; + tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; + tensor transpose_100_perm_0 = const()[name = tensor("transpose_100_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_101_perm_0 = const()[name = tensor("transpose_101_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_101 = transpose(perm = transpose_101_perm_0, x = k_9)[name = tensor("transpose_342")]; + tensor transpose_100 = transpose(perm = transpose_100_perm_0, x = var_911)[name = tensor("transpose_343")]; + tensor matrix_ac_5 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_100, y = transpose_101)[name = tensor("matrix_ac_5")]; + tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_11 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9)[name = tensor("matrix_bd_11")]; + tensor var_937 = add(x = matrix_ac_5, y = matrix_bd_11)[name = tensor("op_937")]; + tensor _inversed_scores_9_y_0 = const()[name = tensor("_inversed_scores_9_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_9 = mul(x = var_937, y = _inversed_scores_9_y_0)[name = tensor("_inversed_scores_9")]; + tensor scores_11 = select(a = var_41, b = _inversed_scores_9, cond = mask_11)[name = tensor("scores_11")]; + tensor var_943 = softmax(axis = var_56, x = scores_11)[name = tensor("op_943")]; + tensor input_145 = select(a = var_40, b = var_943, cond = mask_11)[name = tensor("input_145")]; + tensor x_65_transpose_x_0 = const()[name = tensor("x_65_transpose_x_0"), val = tensor(false)]; + tensor x_65_transpose_y_0 = const()[name = tensor("x_65_transpose_y_0"), val = tensor(false)]; + tensor value_13 = transpose(perm = value_13_perm_0, x = v_5)[name = tensor("transpose_341")]; + tensor x_65 = matmul(transpose_x = x_65_transpose_x_0, transpose_y = x_65_transpose_y_0, x = input_145, y = value_13)[name = tensor("x_65")]; + tensor var_947_perm_0 = const()[name = tensor("op_947_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_948 = const()[name = tensor("op_948"), val = tensor([1, -1, 1024])]; + tensor var_947 = transpose(perm = var_947_perm_0, x = x_65)[name = tensor("transpose_340")]; + tensor input_147 = reshape(shape = var_948, x = var_947)[name = tensor("input_147")]; + tensor input_149 = linear(bias = linear_2_bias_0, weight = module_layers_2_self_attn_linear_out_weight_quantized, x = input_147)[name = tensor("linear_25")]; + tensor input_151 = add(x = input_141, y = input_149)[name = tensor("input_151")]; + tensor x_69_axes_0 = const()[name = tensor("x_69_axes_0"), val = tensor([-1])]; + tensor x_69 = layer_norm(axes = x_69_axes_0, beta = module_layers_2_norm_conv_bias, epsilon = var_38, gamma = module_layers_2_norm_conv_weight, x = input_151)[name = tensor("x_69")]; + tensor input_153_perm_0 = const()[name = tensor("input_153_perm_0"), val = tensor([0, 2, 1])]; + tensor input_155_pad_type_0 = const()[name = tensor("input_155_pad_type_0"), val = tensor("valid")]; + tensor input_155_strides_0 = const()[name = tensor("input_155_strides_0"), val = tensor([1])]; + tensor input_155_pad_0 = const()[name = tensor("input_155_pad_0"), val = tensor([0, 0])]; + tensor input_155_dilations_0 = const()[name = tensor("input_155_dilations_0"), val = tensor([1])]; + tensor input_155_groups_0 = const()[name = tensor("input_155_groups_0"), val = tensor(1)]; + tensor input_153 = transpose(perm = input_153_perm_0, x = x_69)[name = tensor("transpose_339")]; + tensor input_155 = conv(dilations = input_155_dilations_0, groups = input_155_groups_0, pad = input_155_pad_0, pad_type = input_155_pad_type_0, strides = input_155_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_quantized, x = input_153)[name = tensor("input_155")]; + tensor x_71_split_num_splits_0 = const()[name = tensor("x_71_split_num_splits_0"), val = tensor(2)]; + tensor x_71_split_axis_0 = const()[name = tensor("x_71_split_axis_0"), val = tensor(1)]; + tensor x_71_split_0, tensor x_71_split_1 = split(axis = x_71_split_axis_0, num_splits = x_71_split_num_splits_0, x = input_155)[name = tensor("x_71_split")]; + tensor x_71_split_1_sigmoid = sigmoid(x = x_71_split_1)[name = tensor("x_71_split_1_sigmoid")]; + tensor x_71 = mul(x = x_71_split_0, y = x_71_split_1_sigmoid)[name = tensor("x_71")]; + tensor input_157 = select(a = var_40, b = x_71, cond = var_565)[name = tensor("input_157")]; + tensor new_x_11_interleave_0 = const()[name = tensor("new_x_11_interleave_0"), val = tensor(false)]; + tensor new_x_11 = concat(axis = var_56, interleave = new_x_11_interleave_0, values = (cache_11, input_157))[name = tensor("new_x_11")]; + tensor var_986_begin_0 = const()[name = tensor("op_986_begin_0"), val = tensor([0, 0, 2])]; + tensor var_986_end_0 = const()[name = tensor("op_986_end_0"), val = tensor([1, 1024, 10])]; + tensor var_986_end_mask_0 = const()[name = tensor("op_986_end_mask_0"), val = tensor([true, true, true])]; + tensor var_986 = slice_by_index(begin = var_986_begin_0, end = var_986_end_0, end_mask = var_986_end_mask_0, x = new_x_11)[name = tensor("op_986")]; + tensor x_73_pad_type_0 = const()[name = tensor("x_73_pad_type_0"), val = tensor("valid")]; + tensor x_73_groups_0 = const()[name = tensor("x_73_groups_0"), val = tensor(1024)]; + tensor x_73_strides_0 = const()[name = tensor("x_73_strides_0"), val = tensor([1])]; + tensor x_73_pad_0 = const()[name = tensor("x_73_pad_0"), val = tensor([0, 0])]; + tensor x_73_dilations_0 = const()[name = tensor("x_73_dilations_0"), val = tensor([1])]; + tensor x_73 = conv(dilations = x_73_dilations_0, groups = x_73_groups_0, pad = x_73_pad_0, pad_type = x_73_pad_type_0, strides = x_73_strides_0, weight = module_layers_2_conv_depthwise_conv_weight_quantized, x = new_x_11)[name = tensor("x_73")]; + tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; + tensor x_75_axes_0 = const()[name = tensor("x_75_axes_0"), val = tensor([-1])]; + tensor input_159 = transpose(perm = input_159_perm_0, x = x_73)[name = tensor("transpose_338")]; + tensor x_75 = layer_norm(axes = x_75_axes_0, beta = module_layers_2_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_2_conv_batch_norm_weight, x = input_159)[name = tensor("x_75")]; + tensor input_161_perm_0 = const()[name = tensor("input_161_perm_0"), val = tensor([0, 2, 1])]; + tensor input_161 = transpose(perm = input_161_perm_0, x = x_75)[name = tensor("transpose_337")]; + tensor input_163 = silu(x = input_161)[name = tensor("input_163")]; + tensor x_77_pad_type_0 = const()[name = tensor("x_77_pad_type_0"), val = tensor("valid")]; + tensor x_77_strides_0 = const()[name = tensor("x_77_strides_0"), val = tensor([1])]; + tensor x_77_pad_0 = const()[name = tensor("x_77_pad_0"), val = tensor([0, 0])]; + tensor x_77_dilations_0 = const()[name = tensor("x_77_dilations_0"), val = tensor([1])]; + tensor x_77_groups_0 = const()[name = tensor("x_77_groups_0"), val = tensor(1)]; + tensor x_77 = conv(dilations = x_77_dilations_0, groups = x_77_groups_0, pad = x_77_pad_0, pad_type = x_77_pad_type_0, strides = x_77_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_quantized, x = input_163)[name = tensor("x_77")]; + tensor input_165_perm_0 = const()[name = tensor("input_165_perm_0"), val = tensor([0, 2, 1])]; + tensor input_165 = transpose(perm = input_165_perm_0, x = x_77)[name = tensor("transpose_336")]; + tensor input_167 = add(x = input_151, y = input_165)[name = tensor("input_167")]; + tensor input_169_axes_0 = const()[name = tensor("input_169_axes_0"), val = tensor([-1])]; + tensor input_169 = layer_norm(axes = input_169_axes_0, beta = module_layers_2_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_2_norm_feed_forward2_weight, x = input_167)[name = tensor("input_169")]; + tensor input_171 = linear(bias = linear_1_bias_0, weight = module_layers_2_feed_forward2_linear1_weight_quantized, x = input_169)[name = tensor("linear_26")]; + tensor input_173 = silu(x = input_171)[name = tensor("input_173")]; + tensor input_177 = linear(bias = linear_2_bias_0, weight = module_layers_2_feed_forward2_linear2_weight_quantized, x = input_173)[name = tensor("linear_27")]; + tensor var_1027 = const()[name = tensor("op_1027"), val = tensor(0x1p-1)]; + tensor var_1028 = mul(x = input_177, y = var_1027)[name = tensor("op_1028")]; + tensor input_179 = add(x = input_167, y = var_1028)[name = tensor("input_179")]; + tensor input_181_axes_0 = const()[name = tensor("input_181_axes_0"), val = tensor([-1])]; + tensor input_181 = layer_norm(axes = input_181_axes_0, beta = module_layers_2_norm_out_bias, epsilon = var_38, gamma = module_layers_2_norm_out_weight, x = input_179)[name = tensor("input_181")]; + tensor cache_13_begin_0 = const()[name = tensor("cache_13_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_13_end_0 = const()[name = tensor("cache_13_end_0"), val = tensor([4, 1, 70, 1024])]; + tensor cache_13_end_mask_0 = const()[name = tensor("cache_13_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_13_squeeze_mask_0 = const()[name = tensor("cache_13_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_13 = slice_by_index(begin = cache_13_begin_0, end = cache_13_end_0, end_mask = cache_13_end_mask_0, squeeze_mask = cache_13_squeeze_mask_0, x = value_3)[name = tensor("cache_13")]; + tensor cache_15_begin_0 = const()[name = tensor("cache_15_begin_0"), val = tensor([3, 0, 0, 0])]; + tensor cache_15_end_0 = const()[name = tensor("cache_15_end_0"), val = tensor([4, 1, 1024, 8])]; + tensor cache_15_end_mask_0 = const()[name = tensor("cache_15_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_15_squeeze_mask_0 = const()[name = tensor("cache_15_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_15 = slice_by_index(begin = cache_15_begin_0, end = cache_15_end_0, end_mask = cache_15_end_mask_0, squeeze_mask = cache_15_squeeze_mask_0, x = value_5)[name = tensor("cache_15")]; + tensor input_183_axes_0 = const()[name = tensor("input_183_axes_0"), val = tensor([-1])]; + tensor input_183 = layer_norm(axes = input_183_axes_0, beta = module_layers_3_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_3_norm_feed_forward1_weight, x = input_181)[name = tensor("input_183")]; + tensor input_185 = linear(bias = linear_1_bias_0, weight = module_layers_3_feed_forward1_linear1_weight_quantized, x = input_183)[name = tensor("linear_28")]; + tensor input_187 = silu(x = input_185)[name = tensor("input_187")]; + tensor input_191 = linear(bias = linear_2_bias_0, weight = module_layers_3_feed_forward1_linear2_weight_quantized, x = input_187)[name = tensor("linear_29")]; + tensor var_1062 = const()[name = tensor("op_1062"), val = tensor(0x1p-1)]; + tensor var_1063 = mul(x = input_191, y = var_1062)[name = tensor("op_1063")]; + tensor input_193 = add(x = input_181, y = var_1063)[name = tensor("input_193")]; + tensor key_7_axes_0 = const()[name = tensor("key_7_axes_0"), val = tensor([-1])]; + tensor key_7 = layer_norm(axes = key_7_axes_0, beta = module_layers_3_norm_self_att_bias, epsilon = var_38, gamma = module_layers_3_norm_self_att_weight, x = input_193)[name = tensor("key_7")]; + tensor input_195_interleave_0 = const()[name = tensor("input_195_interleave_0"), val = tensor(false)]; + tensor input_195 = concat(axis = var_65, interleave = input_195_interleave_0, values = (cache_13, key_7))[name = tensor("input_195")]; + tensor var_1085_begin_0 = const()[name = tensor("op_1085_begin_0"), val = tensor([0, 2, 0])]; + tensor var_1085_end_0 = const()[name = tensor("op_1085_end_0"), val = tensor([1, 70, 1024])]; + tensor var_1085_end_mask_0 = const()[name = tensor("op_1085_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1085 = slice_by_index(begin = var_1085_begin_0, end = var_1085_end_0, end_mask = var_1085_end_mask_0, x = cache_13)[name = tensor("op_1085")]; + tensor var_1091_interleave_0 = const()[name = tensor("op_1091_interleave_0"), val = tensor(false)]; + tensor var_1091 = concat(axis = var_65, interleave = var_1091_interleave_0, values = (var_1085, key_7))[name = tensor("op_1091")]; + tensor var_1094 = linear(bias = linear_2_bias_0, weight = module_layers_3_self_attn_linear_q_weight_quantized, x = key_7)[name = tensor("linear_30")]; + tensor var_1095 = const()[name = tensor("op_1095"), val = tensor([1, -1, 8, 128])]; + tensor q_19 = reshape(shape = var_1095, x = var_1094)[name = tensor("q_19")]; + tensor var_1098 = linear(bias = linear_2_bias_0, weight = module_layers_3_self_attn_linear_k_weight_quantized, x = input_195)[name = tensor("linear_31")]; + tensor var_1099 = const()[name = tensor("op_1099"), val = tensor([1, -1, 8, 128])]; + tensor k_13 = reshape(shape = var_1099, x = var_1098)[name = tensor("k_13")]; + tensor var_1102 = linear(bias = linear_2_bias_0, weight = module_layers_3_self_attn_linear_v_weight_quantized, x = input_195)[name = tensor("linear_32")]; + tensor var_1103 = const()[name = tensor("op_1103"), val = tensor([1, -1, 8, 128])]; + tensor v_7 = reshape(shape = var_1103, x = var_1102)[name = tensor("v_7")]; + tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_1115 = add(x = q_19, y = module_layers_3_self_attn_pos_bias_u)[name = tensor("op_1115")]; + tensor var_1117 = add(x = q_19, y = module_layers_3_self_attn_pos_bias_v)[name = tensor("op_1117")]; + tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_1119_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1119_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587775360))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587922112))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587921856)))]; + tensor x_85_transpose_x_0 = const()[name = tensor("x_85_transpose_x_0"), val = tensor(false)]; + tensor x_85_transpose_y_0 = const()[name = tensor("x_85_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_7 = transpose(perm = q_with_bias_v_7_perm_0, x = var_1117)[name = tensor("transpose_335")]; + tensor x_85 = matmul(transpose_x = x_85_transpose_x_0, transpose_y = x_85_transpose_y_0, x = q_with_bias_v_7, y = op_1119_quantized)[name = tensor("x_85")]; + tensor const_118 = const()[name = tensor("const_118"), val = tensor(0x0p+0)]; + tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_87_mode_0 = const()[name = tensor("x_87_mode_0"), val = tensor("constant")]; + tensor x_87 = pad(constant_val = const_118, mode = x_87_mode_0, pad = x_87_pad_0, x = x_85)[name = tensor("x_87")]; + tensor var_1127 = const()[name = tensor("op_1127"), val = tensor([1, 8, -1, 2])]; + tensor x_89 = reshape(shape = var_1127, x = x_87)[name = tensor("x_89")]; + tensor var_1131_begin_0 = const()[name = tensor("op_1131_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1131_end_0 = const()[name = tensor("op_1131_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_1131_end_mask_0 = const()[name = tensor("op_1131_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1131 = slice_by_index(begin = var_1131_begin_0, end = var_1131_end_0, end_mask = var_1131_end_mask_0, x = x_89)[name = tensor("op_1131")]; + tensor var_1132 = const()[name = tensor("op_1132"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_13 = reshape(shape = var_1132, x = var_1131)[name = tensor("matrix_bd_13")]; + tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; + tensor transpose_102_perm_0 = const()[name = tensor("transpose_102_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_103_perm_0 = const()[name = tensor("transpose_103_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_103 = transpose(perm = transpose_103_perm_0, x = k_13)[name = tensor("transpose_333")]; + tensor transpose_102 = transpose(perm = transpose_102_perm_0, x = var_1115)[name = tensor("transpose_334")]; + tensor matrix_ac_7 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_102, y = transpose_103)[name = tensor("matrix_ac_7")]; + tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_15 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13)[name = tensor("matrix_bd_15")]; + tensor var_1141 = add(x = matrix_ac_7, y = matrix_bd_15)[name = tensor("op_1141")]; + tensor _inversed_scores_13_y_0 = const()[name = tensor("_inversed_scores_13_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_13 = mul(x = var_1141, y = _inversed_scores_13_y_0)[name = tensor("_inversed_scores_13")]; + tensor scores_15 = select(a = var_41, b = _inversed_scores_13, cond = mask_11)[name = tensor("scores_15")]; + tensor var_1147 = softmax(axis = var_56, x = scores_15)[name = tensor("op_1147")]; + tensor input_197 = select(a = var_40, b = var_1147, cond = mask_11)[name = tensor("input_197")]; + tensor x_91_transpose_x_0 = const()[name = tensor("x_91_transpose_x_0"), val = tensor(false)]; + tensor x_91_transpose_y_0 = const()[name = tensor("x_91_transpose_y_0"), val = tensor(false)]; + tensor value_15 = transpose(perm = value_15_perm_0, x = v_7)[name = tensor("transpose_332")]; + tensor x_91 = matmul(transpose_x = x_91_transpose_x_0, transpose_y = x_91_transpose_y_0, x = input_197, y = value_15)[name = tensor("x_91")]; + tensor var_1151_perm_0 = const()[name = tensor("op_1151_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1152 = const()[name = tensor("op_1152"), val = tensor([1, -1, 1024])]; + tensor var_1151 = transpose(perm = var_1151_perm_0, x = x_91)[name = tensor("transpose_331")]; + tensor input_199 = reshape(shape = var_1152, x = var_1151)[name = tensor("input_199")]; + tensor input_201 = linear(bias = linear_2_bias_0, weight = module_layers_3_self_attn_linear_out_weight_quantized, x = input_199)[name = tensor("linear_34")]; + tensor input_203 = add(x = input_193, y = input_201)[name = tensor("input_203")]; + tensor x_95_axes_0 = const()[name = tensor("x_95_axes_0"), val = tensor([-1])]; + tensor x_95 = layer_norm(axes = x_95_axes_0, beta = module_layers_3_norm_conv_bias, epsilon = var_38, gamma = module_layers_3_norm_conv_weight, x = input_203)[name = tensor("x_95")]; + tensor input_205_perm_0 = const()[name = tensor("input_205_perm_0"), val = tensor([0, 2, 1])]; + tensor input_207_pad_type_0 = const()[name = tensor("input_207_pad_type_0"), val = tensor("valid")]; + tensor input_207_strides_0 = const()[name = tensor("input_207_strides_0"), val = tensor([1])]; + tensor input_207_pad_0 = const()[name = tensor("input_207_pad_0"), val = tensor([0, 0])]; + tensor input_207_dilations_0 = const()[name = tensor("input_207_dilations_0"), val = tensor([1])]; + tensor input_207_groups_0 = const()[name = tensor("input_207_groups_0"), val = tensor(1)]; + tensor input_205 = transpose(perm = input_205_perm_0, x = x_95)[name = tensor("transpose_330")]; + tensor input_207 = conv(dilations = input_207_dilations_0, groups = input_207_groups_0, pad = input_207_pad_0, pad_type = input_207_pad_type_0, strides = input_207_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_quantized, x = input_205)[name = tensor("input_207")]; + tensor x_97_split_num_splits_0 = const()[name = tensor("x_97_split_num_splits_0"), val = tensor(2)]; + tensor x_97_split_axis_0 = const()[name = tensor("x_97_split_axis_0"), val = tensor(1)]; + tensor x_97_split_0, tensor x_97_split_1 = split(axis = x_97_split_axis_0, num_splits = x_97_split_num_splits_0, x = input_207)[name = tensor("x_97_split")]; + tensor x_97_split_1_sigmoid = sigmoid(x = x_97_split_1)[name = tensor("x_97_split_1_sigmoid")]; + tensor x_97 = mul(x = x_97_split_0, y = x_97_split_1_sigmoid)[name = tensor("x_97")]; + tensor input_209 = select(a = var_40, b = x_97, cond = var_565)[name = tensor("input_209")]; + tensor new_x_15_interleave_0 = const()[name = tensor("new_x_15_interleave_0"), val = tensor(false)]; + tensor new_x_15 = concat(axis = var_56, interleave = new_x_15_interleave_0, values = (cache_15, input_209))[name = tensor("new_x_15")]; + tensor var_1190_begin_0 = const()[name = tensor("op_1190_begin_0"), val = tensor([0, 0, 2])]; + tensor var_1190_end_0 = const()[name = tensor("op_1190_end_0"), val = tensor([1, 1024, 10])]; + tensor var_1190_end_mask_0 = const()[name = tensor("op_1190_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1190 = slice_by_index(begin = var_1190_begin_0, end = var_1190_end_0, end_mask = var_1190_end_mask_0, x = new_x_15)[name = tensor("op_1190")]; + tensor x_99_pad_type_0 = const()[name = tensor("x_99_pad_type_0"), val = tensor("valid")]; + tensor x_99_groups_0 = const()[name = tensor("x_99_groups_0"), val = tensor(1024)]; + tensor x_99_strides_0 = const()[name = tensor("x_99_strides_0"), val = tensor([1])]; + tensor x_99_pad_0 = const()[name = tensor("x_99_pad_0"), val = tensor([0, 0])]; + tensor x_99_dilations_0 = const()[name = tensor("x_99_dilations_0"), val = tensor([1])]; + tensor x_99 = conv(dilations = x_99_dilations_0, groups = x_99_groups_0, pad = x_99_pad_0, pad_type = x_99_pad_type_0, strides = x_99_strides_0, weight = module_layers_3_conv_depthwise_conv_weight_quantized, x = new_x_15)[name = tensor("x_99")]; + tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; + tensor x_101_axes_0 = const()[name = tensor("x_101_axes_0"), val = tensor([-1])]; + tensor input_211 = transpose(perm = input_211_perm_0, x = x_99)[name = tensor("transpose_329")]; + tensor x_101 = layer_norm(axes = x_101_axes_0, beta = module_layers_3_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_3_conv_batch_norm_weight, x = input_211)[name = tensor("x_101")]; + tensor input_213_perm_0 = const()[name = tensor("input_213_perm_0"), val = tensor([0, 2, 1])]; + tensor input_213 = transpose(perm = input_213_perm_0, x = x_101)[name = tensor("transpose_328")]; + tensor input_215 = silu(x = input_213)[name = tensor("input_215")]; + tensor x_103_pad_type_0 = const()[name = tensor("x_103_pad_type_0"), val = tensor("valid")]; + tensor x_103_strides_0 = const()[name = tensor("x_103_strides_0"), val = tensor([1])]; + tensor x_103_pad_0 = const()[name = tensor("x_103_pad_0"), val = tensor([0, 0])]; + tensor x_103_dilations_0 = const()[name = tensor("x_103_dilations_0"), val = tensor([1])]; + tensor x_103_groups_0 = const()[name = tensor("x_103_groups_0"), val = tensor(1)]; + tensor x_103 = conv(dilations = x_103_dilations_0, groups = x_103_groups_0, pad = x_103_pad_0, pad_type = x_103_pad_type_0, strides = x_103_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_quantized, x = input_215)[name = tensor("x_103")]; + tensor input_217_perm_0 = const()[name = tensor("input_217_perm_0"), val = tensor([0, 2, 1])]; + tensor input_217 = transpose(perm = input_217_perm_0, x = x_103)[name = tensor("transpose_327")]; + tensor input_219 = add(x = input_203, y = input_217)[name = tensor("input_219")]; + tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; + tensor input_221 = layer_norm(axes = input_221_axes_0, beta = module_layers_3_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_3_norm_feed_forward2_weight, x = input_219)[name = tensor("input_221")]; + tensor input_223 = linear(bias = linear_1_bias_0, weight = module_layers_3_feed_forward2_linear1_weight_quantized, x = input_221)[name = tensor("linear_35")]; + tensor input_225 = silu(x = input_223)[name = tensor("input_225")]; + tensor input_229 = linear(bias = linear_2_bias_0, weight = module_layers_3_feed_forward2_linear2_weight_quantized, x = input_225)[name = tensor("linear_36")]; + tensor var_1231 = const()[name = tensor("op_1231"), val = tensor(0x1p-1)]; + tensor var_1232 = mul(x = input_229, y = var_1231)[name = tensor("op_1232")]; + tensor input_231 = add(x = input_219, y = var_1232)[name = tensor("input_231")]; + tensor input_233_axes_0 = const()[name = tensor("input_233_axes_0"), val = tensor([-1])]; + tensor input_233 = layer_norm(axes = input_233_axes_0, beta = module_layers_3_norm_out_bias, epsilon = var_38, gamma = module_layers_3_norm_out_weight, x = input_231)[name = tensor("input_233")]; + tensor cache_17_begin_0 = const()[name = tensor("cache_17_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_17_end_0 = const()[name = tensor("cache_17_end_0"), val = tensor([5, 1, 70, 1024])]; + tensor cache_17_end_mask_0 = const()[name = tensor("cache_17_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_17_squeeze_mask_0 = const()[name = tensor("cache_17_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_17 = slice_by_index(begin = cache_17_begin_0, end = cache_17_end_0, end_mask = cache_17_end_mask_0, squeeze_mask = cache_17_squeeze_mask_0, x = value_3)[name = tensor("cache_17")]; + tensor cache_19_begin_0 = const()[name = tensor("cache_19_begin_0"), val = tensor([4, 0, 0, 0])]; + tensor cache_19_end_0 = const()[name = tensor("cache_19_end_0"), val = tensor([5, 1, 1024, 8])]; + tensor cache_19_end_mask_0 = const()[name = tensor("cache_19_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_19_squeeze_mask_0 = const()[name = tensor("cache_19_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_19 = slice_by_index(begin = cache_19_begin_0, end = cache_19_end_0, end_mask = cache_19_end_mask_0, squeeze_mask = cache_19_squeeze_mask_0, x = value_5)[name = tensor("cache_19")]; + tensor input_235_axes_0 = const()[name = tensor("input_235_axes_0"), val = tensor([-1])]; + tensor input_235 = layer_norm(axes = input_235_axes_0, beta = module_layers_4_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_4_norm_feed_forward1_weight, x = input_233)[name = tensor("input_235")]; + tensor input_237 = linear(bias = linear_1_bias_0, weight = module_layers_4_feed_forward1_linear1_weight_quantized, x = input_235)[name = tensor("linear_37")]; + tensor input_239 = silu(x = input_237)[name = tensor("input_239")]; + tensor input_243 = linear(bias = linear_2_bias_0, weight = module_layers_4_feed_forward1_linear2_weight_quantized, x = input_239)[name = tensor("linear_38")]; + tensor var_1266 = const()[name = tensor("op_1266"), val = tensor(0x1p-1)]; + tensor var_1267 = mul(x = input_243, y = var_1266)[name = tensor("op_1267")]; + tensor input_245 = add(x = input_233, y = var_1267)[name = tensor("input_245")]; + tensor key_9_axes_0 = const()[name = tensor("key_9_axes_0"), val = tensor([-1])]; + tensor key_9 = layer_norm(axes = key_9_axes_0, beta = module_layers_4_norm_self_att_bias, epsilon = var_38, gamma = module_layers_4_norm_self_att_weight, x = input_245)[name = tensor("key_9")]; + tensor input_247_interleave_0 = const()[name = tensor("input_247_interleave_0"), val = tensor(false)]; + tensor input_247 = concat(axis = var_65, interleave = input_247_interleave_0, values = (cache_17, key_9))[name = tensor("input_247")]; + tensor var_1289_begin_0 = const()[name = tensor("op_1289_begin_0"), val = tensor([0, 2, 0])]; + tensor var_1289_end_0 = const()[name = tensor("op_1289_end_0"), val = tensor([1, 70, 1024])]; + tensor var_1289_end_mask_0 = const()[name = tensor("op_1289_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1289 = slice_by_index(begin = var_1289_begin_0, end = var_1289_end_0, end_mask = var_1289_end_mask_0, x = cache_17)[name = tensor("op_1289")]; + tensor var_1295_interleave_0 = const()[name = tensor("op_1295_interleave_0"), val = tensor(false)]; + tensor var_1295 = concat(axis = var_65, interleave = var_1295_interleave_0, values = (var_1289, key_9))[name = tensor("op_1295")]; + tensor var_1298 = linear(bias = linear_2_bias_0, weight = module_layers_4_self_attn_linear_q_weight_quantized, x = key_9)[name = tensor("linear_39")]; + tensor var_1299 = const()[name = tensor("op_1299"), val = tensor([1, -1, 8, 128])]; + tensor q_25 = reshape(shape = var_1299, x = var_1298)[name = tensor("q_25")]; + tensor var_1302 = linear(bias = linear_2_bias_0, weight = module_layers_4_self_attn_linear_k_weight_quantized, x = input_247)[name = tensor("linear_40")]; + tensor var_1303 = const()[name = tensor("op_1303"), val = tensor([1, -1, 8, 128])]; + tensor k_17 = reshape(shape = var_1303, x = var_1302)[name = tensor("k_17")]; + tensor var_1306 = linear(bias = linear_2_bias_0, weight = module_layers_4_self_attn_linear_v_weight_quantized, x = input_247)[name = tensor("linear_41")]; + tensor var_1307 = const()[name = tensor("op_1307"), val = tensor([1, -1, 8, 128])]; + tensor v_9 = reshape(shape = var_1307, x = var_1306)[name = tensor("v_9")]; + tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_1319 = add(x = q_25, y = module_layers_4_self_attn_pos_bias_u)[name = tensor("op_1319")]; + tensor var_1321 = add(x = q_25, y = module_layers_4_self_attn_pos_bias_v)[name = tensor("op_1321")]; + tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_1323_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1323_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(587922752))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588069504))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588069248)))]; + tensor x_111_transpose_x_0 = const()[name = tensor("x_111_transpose_x_0"), val = tensor(false)]; + tensor x_111_transpose_y_0 = const()[name = tensor("x_111_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_9 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1321)[name = tensor("transpose_326")]; + tensor x_111 = matmul(transpose_x = x_111_transpose_x_0, transpose_y = x_111_transpose_y_0, x = q_with_bias_v_9, y = op_1323_quantized)[name = tensor("x_111")]; + tensor const_131 = const()[name = tensor("const_131"), val = tensor(0x0p+0)]; + tensor x_113_pad_0 = const()[name = tensor("x_113_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_113_mode_0 = const()[name = tensor("x_113_mode_0"), val = tensor("constant")]; + tensor x_113 = pad(constant_val = const_131, mode = x_113_mode_0, pad = x_113_pad_0, x = x_111)[name = tensor("x_113")]; + tensor var_1331 = const()[name = tensor("op_1331"), val = tensor([1, 8, -1, 2])]; + tensor x_115 = reshape(shape = var_1331, x = x_113)[name = tensor("x_115")]; + tensor var_1335_begin_0 = const()[name = tensor("op_1335_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1335_end_0 = const()[name = tensor("op_1335_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_1335_end_mask_0 = const()[name = tensor("op_1335_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1335 = slice_by_index(begin = var_1335_begin_0, end = var_1335_end_0, end_mask = var_1335_end_mask_0, x = x_115)[name = tensor("op_1335")]; + tensor var_1336 = const()[name = tensor("op_1336"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_17 = reshape(shape = var_1336, x = var_1335)[name = tensor("matrix_bd_17")]; + tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; + tensor transpose_104_perm_0 = const()[name = tensor("transpose_104_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_105_perm_0 = const()[name = tensor("transpose_105_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_105 = transpose(perm = transpose_105_perm_0, x = k_17)[name = tensor("transpose_324")]; + tensor transpose_104 = transpose(perm = transpose_104_perm_0, x = var_1319)[name = tensor("transpose_325")]; + tensor matrix_ac_9 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_104, y = transpose_105)[name = tensor("matrix_ac_9")]; + tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_19 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17)[name = tensor("matrix_bd_19")]; + tensor var_1345 = add(x = matrix_ac_9, y = matrix_bd_19)[name = tensor("op_1345")]; + tensor _inversed_scores_17_y_0 = const()[name = tensor("_inversed_scores_17_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_17 = mul(x = var_1345, y = _inversed_scores_17_y_0)[name = tensor("_inversed_scores_17")]; + tensor scores_19 = select(a = var_41, b = _inversed_scores_17, cond = mask_11)[name = tensor("scores_19")]; + tensor var_1351 = softmax(axis = var_56, x = scores_19)[name = tensor("op_1351")]; + tensor input_249 = select(a = var_40, b = var_1351, cond = mask_11)[name = tensor("input_249")]; + tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; + tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; + tensor value_17 = transpose(perm = value_17_perm_0, x = v_9)[name = tensor("transpose_323")]; + tensor x_117 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = input_249, y = value_17)[name = tensor("x_117")]; + tensor var_1355_perm_0 = const()[name = tensor("op_1355_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1356 = const()[name = tensor("op_1356"), val = tensor([1, -1, 1024])]; + tensor var_1355 = transpose(perm = var_1355_perm_0, x = x_117)[name = tensor("transpose_322")]; + tensor input_251 = reshape(shape = var_1356, x = var_1355)[name = tensor("input_251")]; + tensor input_253 = linear(bias = linear_2_bias_0, weight = module_layers_4_self_attn_linear_out_weight_quantized, x = input_251)[name = tensor("linear_43")]; + tensor input_255 = add(x = input_245, y = input_253)[name = tensor("input_255")]; + tensor x_121_axes_0 = const()[name = tensor("x_121_axes_0"), val = tensor([-1])]; + tensor x_121 = layer_norm(axes = x_121_axes_0, beta = module_layers_4_norm_conv_bias, epsilon = var_38, gamma = module_layers_4_norm_conv_weight, x = input_255)[name = tensor("x_121")]; + tensor input_257_perm_0 = const()[name = tensor("input_257_perm_0"), val = tensor([0, 2, 1])]; + tensor input_259_pad_type_0 = const()[name = tensor("input_259_pad_type_0"), val = tensor("valid")]; + tensor input_259_strides_0 = const()[name = tensor("input_259_strides_0"), val = tensor([1])]; + tensor input_259_pad_0 = const()[name = tensor("input_259_pad_0"), val = tensor([0, 0])]; + tensor input_259_dilations_0 = const()[name = tensor("input_259_dilations_0"), val = tensor([1])]; + tensor input_259_groups_0 = const()[name = tensor("input_259_groups_0"), val = tensor(1)]; + tensor input_257 = transpose(perm = input_257_perm_0, x = x_121)[name = tensor("transpose_321")]; + tensor input_259 = conv(dilations = input_259_dilations_0, groups = input_259_groups_0, pad = input_259_pad_0, pad_type = input_259_pad_type_0, strides = input_259_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_quantized, x = input_257)[name = tensor("input_259")]; + tensor x_123_split_num_splits_0 = const()[name = tensor("x_123_split_num_splits_0"), val = tensor(2)]; + tensor x_123_split_axis_0 = const()[name = tensor("x_123_split_axis_0"), val = tensor(1)]; + tensor x_123_split_0, tensor x_123_split_1 = split(axis = x_123_split_axis_0, num_splits = x_123_split_num_splits_0, x = input_259)[name = tensor("x_123_split")]; + tensor x_123_split_1_sigmoid = sigmoid(x = x_123_split_1)[name = tensor("x_123_split_1_sigmoid")]; + tensor x_123 = mul(x = x_123_split_0, y = x_123_split_1_sigmoid)[name = tensor("x_123")]; + tensor input_261 = select(a = var_40, b = x_123, cond = var_565)[name = tensor("input_261")]; + tensor new_x_19_interleave_0 = const()[name = tensor("new_x_19_interleave_0"), val = tensor(false)]; + tensor new_x_19 = concat(axis = var_56, interleave = new_x_19_interleave_0, values = (cache_19, input_261))[name = tensor("new_x_19")]; + tensor var_1394_begin_0 = const()[name = tensor("op_1394_begin_0"), val = tensor([0, 0, 2])]; + tensor var_1394_end_0 = const()[name = tensor("op_1394_end_0"), val = tensor([1, 1024, 10])]; + tensor var_1394_end_mask_0 = const()[name = tensor("op_1394_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1394 = slice_by_index(begin = var_1394_begin_0, end = var_1394_end_0, end_mask = var_1394_end_mask_0, x = new_x_19)[name = tensor("op_1394")]; + tensor x_125_pad_type_0 = const()[name = tensor("x_125_pad_type_0"), val = tensor("valid")]; + tensor x_125_groups_0 = const()[name = tensor("x_125_groups_0"), val = tensor(1024)]; + tensor x_125_strides_0 = const()[name = tensor("x_125_strides_0"), val = tensor([1])]; + tensor x_125_pad_0 = const()[name = tensor("x_125_pad_0"), val = tensor([0, 0])]; + tensor x_125_dilations_0 = const()[name = tensor("x_125_dilations_0"), val = tensor([1])]; + tensor x_125 = conv(dilations = x_125_dilations_0, groups = x_125_groups_0, pad = x_125_pad_0, pad_type = x_125_pad_type_0, strides = x_125_strides_0, weight = module_layers_4_conv_depthwise_conv_weight_quantized, x = new_x_19)[name = tensor("x_125")]; + tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; + tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; + tensor input_263 = transpose(perm = input_263_perm_0, x = x_125)[name = tensor("transpose_320")]; + tensor x_127 = layer_norm(axes = x_127_axes_0, beta = module_layers_4_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_4_conv_batch_norm_weight, x = input_263)[name = tensor("x_127")]; + tensor input_265_perm_0 = const()[name = tensor("input_265_perm_0"), val = tensor([0, 2, 1])]; + tensor input_265 = transpose(perm = input_265_perm_0, x = x_127)[name = tensor("transpose_319")]; + tensor input_267 = silu(x = input_265)[name = tensor("input_267")]; + tensor x_129_pad_type_0 = const()[name = tensor("x_129_pad_type_0"), val = tensor("valid")]; + tensor x_129_strides_0 = const()[name = tensor("x_129_strides_0"), val = tensor([1])]; + tensor x_129_pad_0 = const()[name = tensor("x_129_pad_0"), val = tensor([0, 0])]; + tensor x_129_dilations_0 = const()[name = tensor("x_129_dilations_0"), val = tensor([1])]; + tensor x_129_groups_0 = const()[name = tensor("x_129_groups_0"), val = tensor(1)]; + tensor x_129 = conv(dilations = x_129_dilations_0, groups = x_129_groups_0, pad = x_129_pad_0, pad_type = x_129_pad_type_0, strides = x_129_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_quantized, x = input_267)[name = tensor("x_129")]; + tensor input_269_perm_0 = const()[name = tensor("input_269_perm_0"), val = tensor([0, 2, 1])]; + tensor input_269 = transpose(perm = input_269_perm_0, x = x_129)[name = tensor("transpose_318")]; + tensor input_271 = add(x = input_255, y = input_269)[name = tensor("input_271")]; + tensor input_273_axes_0 = const()[name = tensor("input_273_axes_0"), val = tensor([-1])]; + tensor input_273 = layer_norm(axes = input_273_axes_0, beta = module_layers_4_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_4_norm_feed_forward2_weight, x = input_271)[name = tensor("input_273")]; + tensor input_275 = linear(bias = linear_1_bias_0, weight = module_layers_4_feed_forward2_linear1_weight_quantized, x = input_273)[name = tensor("linear_44")]; + tensor input_277 = silu(x = input_275)[name = tensor("input_277")]; + tensor input_281 = linear(bias = linear_2_bias_0, weight = module_layers_4_feed_forward2_linear2_weight_quantized, x = input_277)[name = tensor("linear_45")]; + tensor var_1435 = const()[name = tensor("op_1435"), val = tensor(0x1p-1)]; + tensor var_1436 = mul(x = input_281, y = var_1435)[name = tensor("op_1436")]; + tensor input_283 = add(x = input_271, y = var_1436)[name = tensor("input_283")]; + tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; + tensor input_285 = layer_norm(axes = input_285_axes_0, beta = module_layers_4_norm_out_bias, epsilon = var_38, gamma = module_layers_4_norm_out_weight, x = input_283)[name = tensor("input_285")]; + tensor cache_21_begin_0 = const()[name = tensor("cache_21_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_21_end_0 = const()[name = tensor("cache_21_end_0"), val = tensor([6, 1, 70, 1024])]; + tensor cache_21_end_mask_0 = const()[name = tensor("cache_21_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_21_squeeze_mask_0 = const()[name = tensor("cache_21_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_21 = slice_by_index(begin = cache_21_begin_0, end = cache_21_end_0, end_mask = cache_21_end_mask_0, squeeze_mask = cache_21_squeeze_mask_0, x = value_3)[name = tensor("cache_21")]; + tensor cache_23_begin_0 = const()[name = tensor("cache_23_begin_0"), val = tensor([5, 0, 0, 0])]; + tensor cache_23_end_0 = const()[name = tensor("cache_23_end_0"), val = tensor([6, 1, 1024, 8])]; + tensor cache_23_end_mask_0 = const()[name = tensor("cache_23_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_23_squeeze_mask_0 = const()[name = tensor("cache_23_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_23 = slice_by_index(begin = cache_23_begin_0, end = cache_23_end_0, end_mask = cache_23_end_mask_0, squeeze_mask = cache_23_squeeze_mask_0, x = value_5)[name = tensor("cache_23")]; + tensor input_287_axes_0 = const()[name = tensor("input_287_axes_0"), val = tensor([-1])]; + tensor input_287 = layer_norm(axes = input_287_axes_0, beta = module_layers_5_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_5_norm_feed_forward1_weight, x = input_285)[name = tensor("input_287")]; + tensor input_289 = linear(bias = linear_1_bias_0, weight = module_layers_5_feed_forward1_linear1_weight_quantized, x = input_287)[name = tensor("linear_46")]; + tensor input_291 = silu(x = input_289)[name = tensor("input_291")]; + tensor input_295 = linear(bias = linear_2_bias_0, weight = module_layers_5_feed_forward1_linear2_weight_quantized, x = input_291)[name = tensor("linear_47")]; + tensor var_1470 = const()[name = tensor("op_1470"), val = tensor(0x1p-1)]; + tensor var_1471 = mul(x = input_295, y = var_1470)[name = tensor("op_1471")]; + tensor input_297 = add(x = input_285, y = var_1471)[name = tensor("input_297")]; + tensor key_11_axes_0 = const()[name = tensor("key_11_axes_0"), val = tensor([-1])]; + tensor key_11 = layer_norm(axes = key_11_axes_0, beta = module_layers_5_norm_self_att_bias, epsilon = var_38, gamma = module_layers_5_norm_self_att_weight, x = input_297)[name = tensor("key_11")]; + tensor input_299_interleave_0 = const()[name = tensor("input_299_interleave_0"), val = tensor(false)]; + tensor input_299 = concat(axis = var_65, interleave = input_299_interleave_0, values = (cache_21, key_11))[name = tensor("input_299")]; + tensor var_1493_begin_0 = const()[name = tensor("op_1493_begin_0"), val = tensor([0, 2, 0])]; + tensor var_1493_end_0 = const()[name = tensor("op_1493_end_0"), val = tensor([1, 70, 1024])]; + tensor var_1493_end_mask_0 = const()[name = tensor("op_1493_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1493 = slice_by_index(begin = var_1493_begin_0, end = var_1493_end_0, end_mask = var_1493_end_mask_0, x = cache_21)[name = tensor("op_1493")]; + tensor var_1499_interleave_0 = const()[name = tensor("op_1499_interleave_0"), val = tensor(false)]; + tensor var_1499 = concat(axis = var_65, interleave = var_1499_interleave_0, values = (var_1493, key_11))[name = tensor("op_1499")]; + tensor var_1502 = linear(bias = linear_2_bias_0, weight = module_layers_5_self_attn_linear_q_weight_quantized, x = key_11)[name = tensor("linear_48")]; + tensor var_1503 = const()[name = tensor("op_1503"), val = tensor([1, -1, 8, 128])]; + tensor q_31 = reshape(shape = var_1503, x = var_1502)[name = tensor("q_31")]; + tensor var_1506 = linear(bias = linear_2_bias_0, weight = module_layers_5_self_attn_linear_k_weight_quantized, x = input_299)[name = tensor("linear_49")]; + tensor var_1507 = const()[name = tensor("op_1507"), val = tensor([1, -1, 8, 128])]; + tensor k_21 = reshape(shape = var_1507, x = var_1506)[name = tensor("k_21")]; + tensor var_1510 = linear(bias = linear_2_bias_0, weight = module_layers_5_self_attn_linear_v_weight_quantized, x = input_299)[name = tensor("linear_50")]; + tensor var_1511 = const()[name = tensor("op_1511"), val = tensor([1, -1, 8, 128])]; + tensor v_11 = reshape(shape = var_1511, x = var_1510)[name = tensor("v_11")]; + tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_1523 = add(x = q_31, y = module_layers_5_self_attn_pos_bias_u)[name = tensor("op_1523")]; + tensor var_1525 = add(x = q_31, y = module_layers_5_self_attn_pos_bias_v)[name = tensor("op_1525")]; + tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_1527_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1527_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588070144))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588216896))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588216640)))]; + tensor x_137_transpose_x_0 = const()[name = tensor("x_137_transpose_x_0"), val = tensor(false)]; + tensor x_137_transpose_y_0 = const()[name = tensor("x_137_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_11 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1525)[name = tensor("transpose_317")]; + tensor x_137 = matmul(transpose_x = x_137_transpose_x_0, transpose_y = x_137_transpose_y_0, x = q_with_bias_v_11, y = op_1527_quantized)[name = tensor("x_137")]; + tensor const_144 = const()[name = tensor("const_144"), val = tensor(0x0p+0)]; + tensor x_139_pad_0 = const()[name = tensor("x_139_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_139_mode_0 = const()[name = tensor("x_139_mode_0"), val = tensor("constant")]; + tensor x_139 = pad(constant_val = const_144, mode = x_139_mode_0, pad = x_139_pad_0, x = x_137)[name = tensor("x_139")]; + tensor var_1535 = const()[name = tensor("op_1535"), val = tensor([1, 8, -1, 2])]; + tensor x_141 = reshape(shape = var_1535, x = x_139)[name = tensor("x_141")]; + tensor var_1539_begin_0 = const()[name = tensor("op_1539_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1539_end_0 = const()[name = tensor("op_1539_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_1539_end_mask_0 = const()[name = tensor("op_1539_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1539 = slice_by_index(begin = var_1539_begin_0, end = var_1539_end_0, end_mask = var_1539_end_mask_0, x = x_141)[name = tensor("op_1539")]; + tensor var_1540 = const()[name = tensor("op_1540"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_21 = reshape(shape = var_1540, x = var_1539)[name = tensor("matrix_bd_21")]; + tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; + tensor transpose_106_perm_0 = const()[name = tensor("transpose_106_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_107_perm_0 = const()[name = tensor("transpose_107_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_107 = transpose(perm = transpose_107_perm_0, x = k_21)[name = tensor("transpose_315")]; + tensor transpose_106 = transpose(perm = transpose_106_perm_0, x = var_1523)[name = tensor("transpose_316")]; + tensor matrix_ac_11 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_106, y = transpose_107)[name = tensor("matrix_ac_11")]; + tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_23 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21)[name = tensor("matrix_bd_23")]; + tensor var_1549 = add(x = matrix_ac_11, y = matrix_bd_23)[name = tensor("op_1549")]; + tensor _inversed_scores_21_y_0 = const()[name = tensor("_inversed_scores_21_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_21 = mul(x = var_1549, y = _inversed_scores_21_y_0)[name = tensor("_inversed_scores_21")]; + tensor scores_23 = select(a = var_41, b = _inversed_scores_21, cond = mask_11)[name = tensor("scores_23")]; + tensor var_1555 = softmax(axis = var_56, x = scores_23)[name = tensor("op_1555")]; + tensor input_301 = select(a = var_40, b = var_1555, cond = mask_11)[name = tensor("input_301")]; + tensor x_143_transpose_x_0 = const()[name = tensor("x_143_transpose_x_0"), val = tensor(false)]; + tensor x_143_transpose_y_0 = const()[name = tensor("x_143_transpose_y_0"), val = tensor(false)]; + tensor value_19 = transpose(perm = value_19_perm_0, x = v_11)[name = tensor("transpose_314")]; + tensor x_143 = matmul(transpose_x = x_143_transpose_x_0, transpose_y = x_143_transpose_y_0, x = input_301, y = value_19)[name = tensor("x_143")]; + tensor var_1559_perm_0 = const()[name = tensor("op_1559_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1560 = const()[name = tensor("op_1560"), val = tensor([1, -1, 1024])]; + tensor var_1559 = transpose(perm = var_1559_perm_0, x = x_143)[name = tensor("transpose_313")]; + tensor input_303 = reshape(shape = var_1560, x = var_1559)[name = tensor("input_303")]; + tensor input_305 = linear(bias = linear_2_bias_0, weight = module_layers_5_self_attn_linear_out_weight_quantized, x = input_303)[name = tensor("linear_52")]; + tensor input_307 = add(x = input_297, y = input_305)[name = tensor("input_307")]; + tensor x_147_axes_0 = const()[name = tensor("x_147_axes_0"), val = tensor([-1])]; + tensor x_147 = layer_norm(axes = x_147_axes_0, beta = module_layers_5_norm_conv_bias, epsilon = var_38, gamma = module_layers_5_norm_conv_weight, x = input_307)[name = tensor("x_147")]; + tensor input_309_perm_0 = const()[name = tensor("input_309_perm_0"), val = tensor([0, 2, 1])]; + tensor input_311_pad_type_0 = const()[name = tensor("input_311_pad_type_0"), val = tensor("valid")]; + tensor input_311_strides_0 = const()[name = tensor("input_311_strides_0"), val = tensor([1])]; + tensor input_311_pad_0 = const()[name = tensor("input_311_pad_0"), val = tensor([0, 0])]; + tensor input_311_dilations_0 = const()[name = tensor("input_311_dilations_0"), val = tensor([1])]; + tensor input_311_groups_0 = const()[name = tensor("input_311_groups_0"), val = tensor(1)]; + tensor input_309 = transpose(perm = input_309_perm_0, x = x_147)[name = tensor("transpose_312")]; + tensor input_311 = conv(dilations = input_311_dilations_0, groups = input_311_groups_0, pad = input_311_pad_0, pad_type = input_311_pad_type_0, strides = input_311_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_quantized, x = input_309)[name = tensor("input_311")]; + tensor x_149_split_num_splits_0 = const()[name = tensor("x_149_split_num_splits_0"), val = tensor(2)]; + tensor x_149_split_axis_0 = const()[name = tensor("x_149_split_axis_0"), val = tensor(1)]; + tensor x_149_split_0, tensor x_149_split_1 = split(axis = x_149_split_axis_0, num_splits = x_149_split_num_splits_0, x = input_311)[name = tensor("x_149_split")]; + tensor x_149_split_1_sigmoid = sigmoid(x = x_149_split_1)[name = tensor("x_149_split_1_sigmoid")]; + tensor x_149 = mul(x = x_149_split_0, y = x_149_split_1_sigmoid)[name = tensor("x_149")]; + tensor input_313 = select(a = var_40, b = x_149, cond = var_565)[name = tensor("input_313")]; + tensor new_x_23_interleave_0 = const()[name = tensor("new_x_23_interleave_0"), val = tensor(false)]; + tensor new_x_23 = concat(axis = var_56, interleave = new_x_23_interleave_0, values = (cache_23, input_313))[name = tensor("new_x_23")]; + tensor var_1598_begin_0 = const()[name = tensor("op_1598_begin_0"), val = tensor([0, 0, 2])]; + tensor var_1598_end_0 = const()[name = tensor("op_1598_end_0"), val = tensor([1, 1024, 10])]; + tensor var_1598_end_mask_0 = const()[name = tensor("op_1598_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1598 = slice_by_index(begin = var_1598_begin_0, end = var_1598_end_0, end_mask = var_1598_end_mask_0, x = new_x_23)[name = tensor("op_1598")]; + tensor x_151_pad_type_0 = const()[name = tensor("x_151_pad_type_0"), val = tensor("valid")]; + tensor x_151_groups_0 = const()[name = tensor("x_151_groups_0"), val = tensor(1024)]; + tensor x_151_strides_0 = const()[name = tensor("x_151_strides_0"), val = tensor([1])]; + tensor x_151_pad_0 = const()[name = tensor("x_151_pad_0"), val = tensor([0, 0])]; + tensor x_151_dilations_0 = const()[name = tensor("x_151_dilations_0"), val = tensor([1])]; + tensor x_151 = conv(dilations = x_151_dilations_0, groups = x_151_groups_0, pad = x_151_pad_0, pad_type = x_151_pad_type_0, strides = x_151_strides_0, weight = module_layers_5_conv_depthwise_conv_weight_quantized, x = new_x_23)[name = tensor("x_151")]; + tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; + tensor x_153_axes_0 = const()[name = tensor("x_153_axes_0"), val = tensor([-1])]; + tensor input_315 = transpose(perm = input_315_perm_0, x = x_151)[name = tensor("transpose_311")]; + tensor x_153 = layer_norm(axes = x_153_axes_0, beta = module_layers_5_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_5_conv_batch_norm_weight, x = input_315)[name = tensor("x_153")]; + tensor input_317_perm_0 = const()[name = tensor("input_317_perm_0"), val = tensor([0, 2, 1])]; + tensor input_317 = transpose(perm = input_317_perm_0, x = x_153)[name = tensor("transpose_310")]; + tensor input_319 = silu(x = input_317)[name = tensor("input_319")]; + tensor x_155_pad_type_0 = const()[name = tensor("x_155_pad_type_0"), val = tensor("valid")]; + tensor x_155_strides_0 = const()[name = tensor("x_155_strides_0"), val = tensor([1])]; + tensor x_155_pad_0 = const()[name = tensor("x_155_pad_0"), val = tensor([0, 0])]; + tensor x_155_dilations_0 = const()[name = tensor("x_155_dilations_0"), val = tensor([1])]; + tensor x_155_groups_0 = const()[name = tensor("x_155_groups_0"), val = tensor(1)]; + tensor x_155 = conv(dilations = x_155_dilations_0, groups = x_155_groups_0, pad = x_155_pad_0, pad_type = x_155_pad_type_0, strides = x_155_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_quantized, x = input_319)[name = tensor("x_155")]; + tensor input_321_perm_0 = const()[name = tensor("input_321_perm_0"), val = tensor([0, 2, 1])]; + tensor input_321 = transpose(perm = input_321_perm_0, x = x_155)[name = tensor("transpose_309")]; + tensor input_323 = add(x = input_307, y = input_321)[name = tensor("input_323")]; + tensor input_325_axes_0 = const()[name = tensor("input_325_axes_0"), val = tensor([-1])]; + tensor input_325 = layer_norm(axes = input_325_axes_0, beta = module_layers_5_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_5_norm_feed_forward2_weight, x = input_323)[name = tensor("input_325")]; + tensor input_327 = linear(bias = linear_1_bias_0, weight = module_layers_5_feed_forward2_linear1_weight_quantized, x = input_325)[name = tensor("linear_53")]; + tensor input_329 = silu(x = input_327)[name = tensor("input_329")]; + tensor input_333 = linear(bias = linear_2_bias_0, weight = module_layers_5_feed_forward2_linear2_weight_quantized, x = input_329)[name = tensor("linear_54")]; + tensor var_1639 = const()[name = tensor("op_1639"), val = tensor(0x1p-1)]; + tensor var_1640 = mul(x = input_333, y = var_1639)[name = tensor("op_1640")]; + tensor input_335 = add(x = input_323, y = var_1640)[name = tensor("input_335")]; + tensor input_337_axes_0 = const()[name = tensor("input_337_axes_0"), val = tensor([-1])]; + tensor input_337 = layer_norm(axes = input_337_axes_0, beta = module_layers_5_norm_out_bias, epsilon = var_38, gamma = module_layers_5_norm_out_weight, x = input_335)[name = tensor("input_337")]; + tensor cache_25_begin_0 = const()[name = tensor("cache_25_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_25_end_0 = const()[name = tensor("cache_25_end_0"), val = tensor([7, 1, 70, 1024])]; + tensor cache_25_end_mask_0 = const()[name = tensor("cache_25_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_25_squeeze_mask_0 = const()[name = tensor("cache_25_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_25 = slice_by_index(begin = cache_25_begin_0, end = cache_25_end_0, end_mask = cache_25_end_mask_0, squeeze_mask = cache_25_squeeze_mask_0, x = value_3)[name = tensor("cache_25")]; + tensor cache_27_begin_0 = const()[name = tensor("cache_27_begin_0"), val = tensor([6, 0, 0, 0])]; + tensor cache_27_end_0 = const()[name = tensor("cache_27_end_0"), val = tensor([7, 1, 1024, 8])]; + tensor cache_27_end_mask_0 = const()[name = tensor("cache_27_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_27_squeeze_mask_0 = const()[name = tensor("cache_27_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_27 = slice_by_index(begin = cache_27_begin_0, end = cache_27_end_0, end_mask = cache_27_end_mask_0, squeeze_mask = cache_27_squeeze_mask_0, x = value_5)[name = tensor("cache_27")]; + tensor input_339_axes_0 = const()[name = tensor("input_339_axes_0"), val = tensor([-1])]; + tensor input_339 = layer_norm(axes = input_339_axes_0, beta = module_layers_6_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_6_norm_feed_forward1_weight, x = input_337)[name = tensor("input_339")]; + tensor input_341 = linear(bias = linear_1_bias_0, weight = module_layers_6_feed_forward1_linear1_weight_quantized, x = input_339)[name = tensor("linear_55")]; + tensor input_343 = silu(x = input_341)[name = tensor("input_343")]; + tensor input_347 = linear(bias = linear_2_bias_0, weight = module_layers_6_feed_forward1_linear2_weight_quantized, x = input_343)[name = tensor("linear_56")]; + tensor var_1674 = const()[name = tensor("op_1674"), val = tensor(0x1p-1)]; + tensor var_1675 = mul(x = input_347, y = var_1674)[name = tensor("op_1675")]; + tensor input_349 = add(x = input_337, y = var_1675)[name = tensor("input_349")]; + tensor key_13_axes_0 = const()[name = tensor("key_13_axes_0"), val = tensor([-1])]; + tensor key_13 = layer_norm(axes = key_13_axes_0, beta = module_layers_6_norm_self_att_bias, epsilon = var_38, gamma = module_layers_6_norm_self_att_weight, x = input_349)[name = tensor("key_13")]; + tensor input_351_interleave_0 = const()[name = tensor("input_351_interleave_0"), val = tensor(false)]; + tensor input_351 = concat(axis = var_65, interleave = input_351_interleave_0, values = (cache_25, key_13))[name = tensor("input_351")]; + tensor var_1697_begin_0 = const()[name = tensor("op_1697_begin_0"), val = tensor([0, 2, 0])]; + tensor var_1697_end_0 = const()[name = tensor("op_1697_end_0"), val = tensor([1, 70, 1024])]; + tensor var_1697_end_mask_0 = const()[name = tensor("op_1697_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1697 = slice_by_index(begin = var_1697_begin_0, end = var_1697_end_0, end_mask = var_1697_end_mask_0, x = cache_25)[name = tensor("op_1697")]; + tensor var_1703_interleave_0 = const()[name = tensor("op_1703_interleave_0"), val = tensor(false)]; + tensor var_1703 = concat(axis = var_65, interleave = var_1703_interleave_0, values = (var_1697, key_13))[name = tensor("op_1703")]; + tensor var_1706 = linear(bias = linear_2_bias_0, weight = module_layers_6_self_attn_linear_q_weight_quantized, x = key_13)[name = tensor("linear_57")]; + tensor var_1707 = const()[name = tensor("op_1707"), val = tensor([1, -1, 8, 128])]; + tensor q_37 = reshape(shape = var_1707, x = var_1706)[name = tensor("q_37")]; + tensor var_1710 = linear(bias = linear_2_bias_0, weight = module_layers_6_self_attn_linear_k_weight_quantized, x = input_351)[name = tensor("linear_58")]; + tensor var_1711 = const()[name = tensor("op_1711"), val = tensor([1, -1, 8, 128])]; + tensor k_25 = reshape(shape = var_1711, x = var_1710)[name = tensor("k_25")]; + tensor var_1714 = linear(bias = linear_2_bias_0, weight = module_layers_6_self_attn_linear_v_weight_quantized, x = input_351)[name = tensor("linear_59")]; + tensor var_1715 = const()[name = tensor("op_1715"), val = tensor([1, -1, 8, 128])]; + tensor v_13 = reshape(shape = var_1715, x = var_1714)[name = tensor("v_13")]; + tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_1727 = add(x = q_37, y = module_layers_6_self_attn_pos_bias_u)[name = tensor("op_1727")]; + tensor var_1729 = add(x = q_37, y = module_layers_6_self_attn_pos_bias_v)[name = tensor("op_1729")]; + tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_1731_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1731_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588217536))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588364288))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588364032)))]; + tensor x_163_transpose_x_0 = const()[name = tensor("x_163_transpose_x_0"), val = tensor(false)]; + tensor x_163_transpose_y_0 = const()[name = tensor("x_163_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_13 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1729)[name = tensor("transpose_308")]; + tensor x_163 = matmul(transpose_x = x_163_transpose_x_0, transpose_y = x_163_transpose_y_0, x = q_with_bias_v_13, y = op_1731_quantized)[name = tensor("x_163")]; + tensor const_157 = const()[name = tensor("const_157"), val = tensor(0x0p+0)]; + tensor x_165_pad_0 = const()[name = tensor("x_165_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_165_mode_0 = const()[name = tensor("x_165_mode_0"), val = tensor("constant")]; + tensor x_165 = pad(constant_val = const_157, mode = x_165_mode_0, pad = x_165_pad_0, x = x_163)[name = tensor("x_165")]; + tensor var_1739 = const()[name = tensor("op_1739"), val = tensor([1, 8, -1, 2])]; + tensor x_167 = reshape(shape = var_1739, x = x_165)[name = tensor("x_167")]; + tensor var_1743_begin_0 = const()[name = tensor("op_1743_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1743_end_0 = const()[name = tensor("op_1743_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_1743_end_mask_0 = const()[name = tensor("op_1743_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1743 = slice_by_index(begin = var_1743_begin_0, end = var_1743_end_0, end_mask = var_1743_end_mask_0, x = x_167)[name = tensor("op_1743")]; + tensor var_1744 = const()[name = tensor("op_1744"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_25 = reshape(shape = var_1744, x = var_1743)[name = tensor("matrix_bd_25")]; + tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; + tensor transpose_108_perm_0 = const()[name = tensor("transpose_108_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_109_perm_0 = const()[name = tensor("transpose_109_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_109 = transpose(perm = transpose_109_perm_0, x = k_25)[name = tensor("transpose_306")]; + tensor transpose_108 = transpose(perm = transpose_108_perm_0, x = var_1727)[name = tensor("transpose_307")]; + tensor matrix_ac_13 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_108, y = transpose_109)[name = tensor("matrix_ac_13")]; + tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_27 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25)[name = tensor("matrix_bd_27")]; + tensor var_1753 = add(x = matrix_ac_13, y = matrix_bd_27)[name = tensor("op_1753")]; + tensor _inversed_scores_25_y_0 = const()[name = tensor("_inversed_scores_25_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_25 = mul(x = var_1753, y = _inversed_scores_25_y_0)[name = tensor("_inversed_scores_25")]; + tensor scores_27 = select(a = var_41, b = _inversed_scores_25, cond = mask_11)[name = tensor("scores_27")]; + tensor var_1759 = softmax(axis = var_56, x = scores_27)[name = tensor("op_1759")]; + tensor input_353 = select(a = var_40, b = var_1759, cond = mask_11)[name = tensor("input_353")]; + tensor x_169_transpose_x_0 = const()[name = tensor("x_169_transpose_x_0"), val = tensor(false)]; + tensor x_169_transpose_y_0 = const()[name = tensor("x_169_transpose_y_0"), val = tensor(false)]; + tensor value_21 = transpose(perm = value_21_perm_0, x = v_13)[name = tensor("transpose_305")]; + tensor x_169 = matmul(transpose_x = x_169_transpose_x_0, transpose_y = x_169_transpose_y_0, x = input_353, y = value_21)[name = tensor("x_169")]; + tensor var_1763_perm_0 = const()[name = tensor("op_1763_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1764 = const()[name = tensor("op_1764"), val = tensor([1, -1, 1024])]; + tensor var_1763 = transpose(perm = var_1763_perm_0, x = x_169)[name = tensor("transpose_304")]; + tensor input_355 = reshape(shape = var_1764, x = var_1763)[name = tensor("input_355")]; + tensor input_357 = linear(bias = linear_2_bias_0, weight = module_layers_6_self_attn_linear_out_weight_quantized, x = input_355)[name = tensor("linear_61")]; + tensor input_359 = add(x = input_349, y = input_357)[name = tensor("input_359")]; + tensor x_173_axes_0 = const()[name = tensor("x_173_axes_0"), val = tensor([-1])]; + tensor x_173 = layer_norm(axes = x_173_axes_0, beta = module_layers_6_norm_conv_bias, epsilon = var_38, gamma = module_layers_6_norm_conv_weight, x = input_359)[name = tensor("x_173")]; + tensor input_361_perm_0 = const()[name = tensor("input_361_perm_0"), val = tensor([0, 2, 1])]; + tensor input_363_pad_type_0 = const()[name = tensor("input_363_pad_type_0"), val = tensor("valid")]; + tensor input_363_strides_0 = const()[name = tensor("input_363_strides_0"), val = tensor([1])]; + tensor input_363_pad_0 = const()[name = tensor("input_363_pad_0"), val = tensor([0, 0])]; + tensor input_363_dilations_0 = const()[name = tensor("input_363_dilations_0"), val = tensor([1])]; + tensor input_363_groups_0 = const()[name = tensor("input_363_groups_0"), val = tensor(1)]; + tensor input_361 = transpose(perm = input_361_perm_0, x = x_173)[name = tensor("transpose_303")]; + tensor input_363 = conv(dilations = input_363_dilations_0, groups = input_363_groups_0, pad = input_363_pad_0, pad_type = input_363_pad_type_0, strides = input_363_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_quantized, x = input_361)[name = tensor("input_363")]; + tensor x_175_split_num_splits_0 = const()[name = tensor("x_175_split_num_splits_0"), val = tensor(2)]; + tensor x_175_split_axis_0 = const()[name = tensor("x_175_split_axis_0"), val = tensor(1)]; + tensor x_175_split_0, tensor x_175_split_1 = split(axis = x_175_split_axis_0, num_splits = x_175_split_num_splits_0, x = input_363)[name = tensor("x_175_split")]; + tensor x_175_split_1_sigmoid = sigmoid(x = x_175_split_1)[name = tensor("x_175_split_1_sigmoid")]; + tensor x_175 = mul(x = x_175_split_0, y = x_175_split_1_sigmoid)[name = tensor("x_175")]; + tensor input_365 = select(a = var_40, b = x_175, cond = var_565)[name = tensor("input_365")]; + tensor new_x_27_interleave_0 = const()[name = tensor("new_x_27_interleave_0"), val = tensor(false)]; + tensor new_x_27 = concat(axis = var_56, interleave = new_x_27_interleave_0, values = (cache_27, input_365))[name = tensor("new_x_27")]; + tensor var_1802_begin_0 = const()[name = tensor("op_1802_begin_0"), val = tensor([0, 0, 2])]; + tensor var_1802_end_0 = const()[name = tensor("op_1802_end_0"), val = tensor([1, 1024, 10])]; + tensor var_1802_end_mask_0 = const()[name = tensor("op_1802_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1802 = slice_by_index(begin = var_1802_begin_0, end = var_1802_end_0, end_mask = var_1802_end_mask_0, x = new_x_27)[name = tensor("op_1802")]; + tensor x_177_pad_type_0 = const()[name = tensor("x_177_pad_type_0"), val = tensor("valid")]; + tensor x_177_groups_0 = const()[name = tensor("x_177_groups_0"), val = tensor(1024)]; + tensor x_177_strides_0 = const()[name = tensor("x_177_strides_0"), val = tensor([1])]; + tensor x_177_pad_0 = const()[name = tensor("x_177_pad_0"), val = tensor([0, 0])]; + tensor x_177_dilations_0 = const()[name = tensor("x_177_dilations_0"), val = tensor([1])]; + tensor x_177 = conv(dilations = x_177_dilations_0, groups = x_177_groups_0, pad = x_177_pad_0, pad_type = x_177_pad_type_0, strides = x_177_strides_0, weight = module_layers_6_conv_depthwise_conv_weight_quantized, x = new_x_27)[name = tensor("x_177")]; + tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; + tensor x_179_axes_0 = const()[name = tensor("x_179_axes_0"), val = tensor([-1])]; + tensor input_367 = transpose(perm = input_367_perm_0, x = x_177)[name = tensor("transpose_302")]; + tensor x_179 = layer_norm(axes = x_179_axes_0, beta = module_layers_6_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_6_conv_batch_norm_weight, x = input_367)[name = tensor("x_179")]; + tensor input_369_perm_0 = const()[name = tensor("input_369_perm_0"), val = tensor([0, 2, 1])]; + tensor input_369 = transpose(perm = input_369_perm_0, x = x_179)[name = tensor("transpose_301")]; + tensor input_371 = silu(x = input_369)[name = tensor("input_371")]; + tensor x_181_pad_type_0 = const()[name = tensor("x_181_pad_type_0"), val = tensor("valid")]; + tensor x_181_strides_0 = const()[name = tensor("x_181_strides_0"), val = tensor([1])]; + tensor x_181_pad_0 = const()[name = tensor("x_181_pad_0"), val = tensor([0, 0])]; + tensor x_181_dilations_0 = const()[name = tensor("x_181_dilations_0"), val = tensor([1])]; + tensor x_181_groups_0 = const()[name = tensor("x_181_groups_0"), val = tensor(1)]; + tensor x_181 = conv(dilations = x_181_dilations_0, groups = x_181_groups_0, pad = x_181_pad_0, pad_type = x_181_pad_type_0, strides = x_181_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_quantized, x = input_371)[name = tensor("x_181")]; + tensor input_373_perm_0 = const()[name = tensor("input_373_perm_0"), val = tensor([0, 2, 1])]; + tensor input_373 = transpose(perm = input_373_perm_0, x = x_181)[name = tensor("transpose_300")]; + tensor input_375 = add(x = input_359, y = input_373)[name = tensor("input_375")]; + tensor input_377_axes_0 = const()[name = tensor("input_377_axes_0"), val = tensor([-1])]; + tensor input_377 = layer_norm(axes = input_377_axes_0, beta = module_layers_6_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_6_norm_feed_forward2_weight, x = input_375)[name = tensor("input_377")]; + tensor input_379 = linear(bias = linear_1_bias_0, weight = module_layers_6_feed_forward2_linear1_weight_quantized, x = input_377)[name = tensor("linear_62")]; + tensor input_381 = silu(x = input_379)[name = tensor("input_381")]; + tensor input_385 = linear(bias = linear_2_bias_0, weight = module_layers_6_feed_forward2_linear2_weight_quantized, x = input_381)[name = tensor("linear_63")]; + tensor var_1843 = const()[name = tensor("op_1843"), val = tensor(0x1p-1)]; + tensor var_1844 = mul(x = input_385, y = var_1843)[name = tensor("op_1844")]; + tensor input_387 = add(x = input_375, y = var_1844)[name = tensor("input_387")]; + tensor input_389_axes_0 = const()[name = tensor("input_389_axes_0"), val = tensor([-1])]; + tensor input_389 = layer_norm(axes = input_389_axes_0, beta = module_layers_6_norm_out_bias, epsilon = var_38, gamma = module_layers_6_norm_out_weight, x = input_387)[name = tensor("input_389")]; + tensor cache_29_begin_0 = const()[name = tensor("cache_29_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_29_end_0 = const()[name = tensor("cache_29_end_0"), val = tensor([8, 1, 70, 1024])]; + tensor cache_29_end_mask_0 = const()[name = tensor("cache_29_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_29_squeeze_mask_0 = const()[name = tensor("cache_29_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_29 = slice_by_index(begin = cache_29_begin_0, end = cache_29_end_0, end_mask = cache_29_end_mask_0, squeeze_mask = cache_29_squeeze_mask_0, x = value_3)[name = tensor("cache_29")]; + tensor cache_31_begin_0 = const()[name = tensor("cache_31_begin_0"), val = tensor([7, 0, 0, 0])]; + tensor cache_31_end_0 = const()[name = tensor("cache_31_end_0"), val = tensor([8, 1, 1024, 8])]; + tensor cache_31_end_mask_0 = const()[name = tensor("cache_31_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_31_squeeze_mask_0 = const()[name = tensor("cache_31_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_31 = slice_by_index(begin = cache_31_begin_0, end = cache_31_end_0, end_mask = cache_31_end_mask_0, squeeze_mask = cache_31_squeeze_mask_0, x = value_5)[name = tensor("cache_31")]; + tensor input_391_axes_0 = const()[name = tensor("input_391_axes_0"), val = tensor([-1])]; + tensor input_391 = layer_norm(axes = input_391_axes_0, beta = module_layers_7_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_7_norm_feed_forward1_weight, x = input_389)[name = tensor("input_391")]; + tensor input_393 = linear(bias = linear_1_bias_0, weight = module_layers_7_feed_forward1_linear1_weight_quantized, x = input_391)[name = tensor("linear_64")]; + tensor input_395 = silu(x = input_393)[name = tensor("input_395")]; + tensor input_399 = linear(bias = linear_2_bias_0, weight = module_layers_7_feed_forward1_linear2_weight_quantized, x = input_395)[name = tensor("linear_65")]; + tensor var_1878 = const()[name = tensor("op_1878"), val = tensor(0x1p-1)]; + tensor var_1879 = mul(x = input_399, y = var_1878)[name = tensor("op_1879")]; + tensor input_401 = add(x = input_389, y = var_1879)[name = tensor("input_401")]; + tensor key_15_axes_0 = const()[name = tensor("key_15_axes_0"), val = tensor([-1])]; + tensor key_15 = layer_norm(axes = key_15_axes_0, beta = module_layers_7_norm_self_att_bias, epsilon = var_38, gamma = module_layers_7_norm_self_att_weight, x = input_401)[name = tensor("key_15")]; + tensor input_403_interleave_0 = const()[name = tensor("input_403_interleave_0"), val = tensor(false)]; + tensor input_403 = concat(axis = var_65, interleave = input_403_interleave_0, values = (cache_29, key_15))[name = tensor("input_403")]; + tensor var_1901_begin_0 = const()[name = tensor("op_1901_begin_0"), val = tensor([0, 2, 0])]; + tensor var_1901_end_0 = const()[name = tensor("op_1901_end_0"), val = tensor([1, 70, 1024])]; + tensor var_1901_end_mask_0 = const()[name = tensor("op_1901_end_mask_0"), val = tensor([true, true, true])]; + tensor var_1901 = slice_by_index(begin = var_1901_begin_0, end = var_1901_end_0, end_mask = var_1901_end_mask_0, x = cache_29)[name = tensor("op_1901")]; + tensor var_1907_interleave_0 = const()[name = tensor("op_1907_interleave_0"), val = tensor(false)]; + tensor var_1907 = concat(axis = var_65, interleave = var_1907_interleave_0, values = (var_1901, key_15))[name = tensor("op_1907")]; + tensor var_1910 = linear(bias = linear_2_bias_0, weight = module_layers_7_self_attn_linear_q_weight_quantized, x = key_15)[name = tensor("linear_66")]; + tensor var_1911 = const()[name = tensor("op_1911"), val = tensor([1, -1, 8, 128])]; + tensor q_43 = reshape(shape = var_1911, x = var_1910)[name = tensor("q_43")]; + tensor var_1914 = linear(bias = linear_2_bias_0, weight = module_layers_7_self_attn_linear_k_weight_quantized, x = input_403)[name = tensor("linear_67")]; + tensor var_1915 = const()[name = tensor("op_1915"), val = tensor([1, -1, 8, 128])]; + tensor k_29 = reshape(shape = var_1915, x = var_1914)[name = tensor("k_29")]; + tensor var_1918 = linear(bias = linear_2_bias_0, weight = module_layers_7_self_attn_linear_v_weight_quantized, x = input_403)[name = tensor("linear_68")]; + tensor var_1919 = const()[name = tensor("op_1919"), val = tensor([1, -1, 8, 128])]; + tensor v_15 = reshape(shape = var_1919, x = var_1918)[name = tensor("v_15")]; + tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_1931 = add(x = q_43, y = module_layers_7_self_attn_pos_bias_u)[name = tensor("op_1931")]; + tensor var_1933 = add(x = q_43, y = module_layers_7_self_attn_pos_bias_v)[name = tensor("op_1933")]; + tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_1935_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_1935_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588364928))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588511680))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588511424)))]; + tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; + tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_15 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1933)[name = tensor("transpose_299")]; + tensor x_189 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = q_with_bias_v_15, y = op_1935_quantized)[name = tensor("x_189")]; + tensor const_170 = const()[name = tensor("const_170"), val = tensor(0x0p+0)]; + tensor x_191_pad_0 = const()[name = tensor("x_191_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_191_mode_0 = const()[name = tensor("x_191_mode_0"), val = tensor("constant")]; + tensor x_191 = pad(constant_val = const_170, mode = x_191_mode_0, pad = x_191_pad_0, x = x_189)[name = tensor("x_191")]; + tensor var_1943 = const()[name = tensor("op_1943"), val = tensor([1, 8, -1, 2])]; + tensor x_193 = reshape(shape = var_1943, x = x_191)[name = tensor("x_193")]; + tensor var_1947_begin_0 = const()[name = tensor("op_1947_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_1947_end_0 = const()[name = tensor("op_1947_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_1947_end_mask_0 = const()[name = tensor("op_1947_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_1947 = slice_by_index(begin = var_1947_begin_0, end = var_1947_end_0, end_mask = var_1947_end_mask_0, x = x_193)[name = tensor("op_1947")]; + tensor var_1948 = const()[name = tensor("op_1948"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_29 = reshape(shape = var_1948, x = var_1947)[name = tensor("matrix_bd_29")]; + tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; + tensor transpose_110_perm_0 = const()[name = tensor("transpose_110_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_111_perm_0 = const()[name = tensor("transpose_111_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_111 = transpose(perm = transpose_111_perm_0, x = k_29)[name = tensor("transpose_297")]; + tensor transpose_110 = transpose(perm = transpose_110_perm_0, x = var_1931)[name = tensor("transpose_298")]; + tensor matrix_ac_15 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_110, y = transpose_111)[name = tensor("matrix_ac_15")]; + tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_31 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29)[name = tensor("matrix_bd_31")]; + tensor var_1957 = add(x = matrix_ac_15, y = matrix_bd_31)[name = tensor("op_1957")]; + tensor _inversed_scores_29_y_0 = const()[name = tensor("_inversed_scores_29_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_29 = mul(x = var_1957, y = _inversed_scores_29_y_0)[name = tensor("_inversed_scores_29")]; + tensor scores_31 = select(a = var_41, b = _inversed_scores_29, cond = mask_11)[name = tensor("scores_31")]; + tensor var_1963 = softmax(axis = var_56, x = scores_31)[name = tensor("op_1963")]; + tensor input_405 = select(a = var_40, b = var_1963, cond = mask_11)[name = tensor("input_405")]; + tensor x_195_transpose_x_0 = const()[name = tensor("x_195_transpose_x_0"), val = tensor(false)]; + tensor x_195_transpose_y_0 = const()[name = tensor("x_195_transpose_y_0"), val = tensor(false)]; + tensor value_23 = transpose(perm = value_23_perm_0, x = v_15)[name = tensor("transpose_296")]; + tensor x_195 = matmul(transpose_x = x_195_transpose_x_0, transpose_y = x_195_transpose_y_0, x = input_405, y = value_23)[name = tensor("x_195")]; + tensor var_1967_perm_0 = const()[name = tensor("op_1967_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1968 = const()[name = tensor("op_1968"), val = tensor([1, -1, 1024])]; + tensor var_1967 = transpose(perm = var_1967_perm_0, x = x_195)[name = tensor("transpose_295")]; + tensor input_407 = reshape(shape = var_1968, x = var_1967)[name = tensor("input_407")]; + tensor input_409 = linear(bias = linear_2_bias_0, weight = module_layers_7_self_attn_linear_out_weight_quantized, x = input_407)[name = tensor("linear_70")]; + tensor input_411 = add(x = input_401, y = input_409)[name = tensor("input_411")]; + tensor x_199_axes_0 = const()[name = tensor("x_199_axes_0"), val = tensor([-1])]; + tensor x_199 = layer_norm(axes = x_199_axes_0, beta = module_layers_7_norm_conv_bias, epsilon = var_38, gamma = module_layers_7_norm_conv_weight, x = input_411)[name = tensor("x_199")]; + tensor input_413_perm_0 = const()[name = tensor("input_413_perm_0"), val = tensor([0, 2, 1])]; + tensor input_415_pad_type_0 = const()[name = tensor("input_415_pad_type_0"), val = tensor("valid")]; + tensor input_415_strides_0 = const()[name = tensor("input_415_strides_0"), val = tensor([1])]; + tensor input_415_pad_0 = const()[name = tensor("input_415_pad_0"), val = tensor([0, 0])]; + tensor input_415_dilations_0 = const()[name = tensor("input_415_dilations_0"), val = tensor([1])]; + tensor input_415_groups_0 = const()[name = tensor("input_415_groups_0"), val = tensor(1)]; + tensor input_413 = transpose(perm = input_413_perm_0, x = x_199)[name = tensor("transpose_294")]; + tensor input_415 = conv(dilations = input_415_dilations_0, groups = input_415_groups_0, pad = input_415_pad_0, pad_type = input_415_pad_type_0, strides = input_415_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_quantized, x = input_413)[name = tensor("input_415")]; + tensor x_201_split_num_splits_0 = const()[name = tensor("x_201_split_num_splits_0"), val = tensor(2)]; + tensor x_201_split_axis_0 = const()[name = tensor("x_201_split_axis_0"), val = tensor(1)]; + tensor x_201_split_0, tensor x_201_split_1 = split(axis = x_201_split_axis_0, num_splits = x_201_split_num_splits_0, x = input_415)[name = tensor("x_201_split")]; + tensor x_201_split_1_sigmoid = sigmoid(x = x_201_split_1)[name = tensor("x_201_split_1_sigmoid")]; + tensor x_201 = mul(x = x_201_split_0, y = x_201_split_1_sigmoid)[name = tensor("x_201")]; + tensor input_417 = select(a = var_40, b = x_201, cond = var_565)[name = tensor("input_417")]; + tensor new_x_31_interleave_0 = const()[name = tensor("new_x_31_interleave_0"), val = tensor(false)]; + tensor new_x_31 = concat(axis = var_56, interleave = new_x_31_interleave_0, values = (cache_31, input_417))[name = tensor("new_x_31")]; + tensor var_2006_begin_0 = const()[name = tensor("op_2006_begin_0"), val = tensor([0, 0, 2])]; + tensor var_2006_end_0 = const()[name = tensor("op_2006_end_0"), val = tensor([1, 1024, 10])]; + tensor var_2006_end_mask_0 = const()[name = tensor("op_2006_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2006 = slice_by_index(begin = var_2006_begin_0, end = var_2006_end_0, end_mask = var_2006_end_mask_0, x = new_x_31)[name = tensor("op_2006")]; + tensor x_203_pad_type_0 = const()[name = tensor("x_203_pad_type_0"), val = tensor("valid")]; + tensor x_203_groups_0 = const()[name = tensor("x_203_groups_0"), val = tensor(1024)]; + tensor x_203_strides_0 = const()[name = tensor("x_203_strides_0"), val = tensor([1])]; + tensor x_203_pad_0 = const()[name = tensor("x_203_pad_0"), val = tensor([0, 0])]; + tensor x_203_dilations_0 = const()[name = tensor("x_203_dilations_0"), val = tensor([1])]; + tensor x_203 = conv(dilations = x_203_dilations_0, groups = x_203_groups_0, pad = x_203_pad_0, pad_type = x_203_pad_type_0, strides = x_203_strides_0, weight = module_layers_7_conv_depthwise_conv_weight_quantized, x = new_x_31)[name = tensor("x_203")]; + tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; + tensor x_205_axes_0 = const()[name = tensor("x_205_axes_0"), val = tensor([-1])]; + tensor input_419 = transpose(perm = input_419_perm_0, x = x_203)[name = tensor("transpose_293")]; + tensor x_205 = layer_norm(axes = x_205_axes_0, beta = module_layers_7_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_7_conv_batch_norm_weight, x = input_419)[name = tensor("x_205")]; + tensor input_421_perm_0 = const()[name = tensor("input_421_perm_0"), val = tensor([0, 2, 1])]; + tensor input_421 = transpose(perm = input_421_perm_0, x = x_205)[name = tensor("transpose_292")]; + tensor input_423 = silu(x = input_421)[name = tensor("input_423")]; + tensor x_207_pad_type_0 = const()[name = tensor("x_207_pad_type_0"), val = tensor("valid")]; + tensor x_207_strides_0 = const()[name = tensor("x_207_strides_0"), val = tensor([1])]; + tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0])]; + tensor x_207_dilations_0 = const()[name = tensor("x_207_dilations_0"), val = tensor([1])]; + tensor x_207_groups_0 = const()[name = tensor("x_207_groups_0"), val = tensor(1)]; + tensor x_207 = conv(dilations = x_207_dilations_0, groups = x_207_groups_0, pad = x_207_pad_0, pad_type = x_207_pad_type_0, strides = x_207_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_quantized, x = input_423)[name = tensor("x_207")]; + tensor input_425_perm_0 = const()[name = tensor("input_425_perm_0"), val = tensor([0, 2, 1])]; + tensor input_425 = transpose(perm = input_425_perm_0, x = x_207)[name = tensor("transpose_291")]; + tensor input_427 = add(x = input_411, y = input_425)[name = tensor("input_427")]; + tensor input_429_axes_0 = const()[name = tensor("input_429_axes_0"), val = tensor([-1])]; + tensor input_429 = layer_norm(axes = input_429_axes_0, beta = module_layers_7_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_7_norm_feed_forward2_weight, x = input_427)[name = tensor("input_429")]; + tensor input_431 = linear(bias = linear_1_bias_0, weight = module_layers_7_feed_forward2_linear1_weight_quantized, x = input_429)[name = tensor("linear_71")]; + tensor input_433 = silu(x = input_431)[name = tensor("input_433")]; + tensor input_437 = linear(bias = linear_2_bias_0, weight = module_layers_7_feed_forward2_linear2_weight_quantized, x = input_433)[name = tensor("linear_72")]; + tensor var_2047 = const()[name = tensor("op_2047"), val = tensor(0x1p-1)]; + tensor var_2048 = mul(x = input_437, y = var_2047)[name = tensor("op_2048")]; + tensor input_439 = add(x = input_427, y = var_2048)[name = tensor("input_439")]; + tensor input_441_axes_0 = const()[name = tensor("input_441_axes_0"), val = tensor([-1])]; + tensor input_441 = layer_norm(axes = input_441_axes_0, beta = module_layers_7_norm_out_bias, epsilon = var_38, gamma = module_layers_7_norm_out_weight, x = input_439)[name = tensor("input_441")]; + tensor cache_33_begin_0 = const()[name = tensor("cache_33_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_33_end_0 = const()[name = tensor("cache_33_end_0"), val = tensor([9, 1, 70, 1024])]; + tensor cache_33_end_mask_0 = const()[name = tensor("cache_33_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_33_squeeze_mask_0 = const()[name = tensor("cache_33_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_33 = slice_by_index(begin = cache_33_begin_0, end = cache_33_end_0, end_mask = cache_33_end_mask_0, squeeze_mask = cache_33_squeeze_mask_0, x = value_3)[name = tensor("cache_33")]; + tensor cache_35_begin_0 = const()[name = tensor("cache_35_begin_0"), val = tensor([8, 0, 0, 0])]; + tensor cache_35_end_0 = const()[name = tensor("cache_35_end_0"), val = tensor([9, 1, 1024, 8])]; + tensor cache_35_end_mask_0 = const()[name = tensor("cache_35_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_35_squeeze_mask_0 = const()[name = tensor("cache_35_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_35 = slice_by_index(begin = cache_35_begin_0, end = cache_35_end_0, end_mask = cache_35_end_mask_0, squeeze_mask = cache_35_squeeze_mask_0, x = value_5)[name = tensor("cache_35")]; + tensor input_443_axes_0 = const()[name = tensor("input_443_axes_0"), val = tensor([-1])]; + tensor input_443 = layer_norm(axes = input_443_axes_0, beta = module_layers_8_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_8_norm_feed_forward1_weight, x = input_441)[name = tensor("input_443")]; + tensor input_445 = linear(bias = linear_1_bias_0, weight = module_layers_8_feed_forward1_linear1_weight_quantized, x = input_443)[name = tensor("linear_73")]; + tensor input_447 = silu(x = input_445)[name = tensor("input_447")]; + tensor input_451 = linear(bias = linear_2_bias_0, weight = module_layers_8_feed_forward1_linear2_weight_quantized, x = input_447)[name = tensor("linear_74")]; + tensor var_2082 = const()[name = tensor("op_2082"), val = tensor(0x1p-1)]; + tensor var_2083 = mul(x = input_451, y = var_2082)[name = tensor("op_2083")]; + tensor input_453 = add(x = input_441, y = var_2083)[name = tensor("input_453")]; + tensor key_17_axes_0 = const()[name = tensor("key_17_axes_0"), val = tensor([-1])]; + tensor key_17 = layer_norm(axes = key_17_axes_0, beta = module_layers_8_norm_self_att_bias, epsilon = var_38, gamma = module_layers_8_norm_self_att_weight, x = input_453)[name = tensor("key_17")]; + tensor input_455_interleave_0 = const()[name = tensor("input_455_interleave_0"), val = tensor(false)]; + tensor input_455 = concat(axis = var_65, interleave = input_455_interleave_0, values = (cache_33, key_17))[name = tensor("input_455")]; + tensor var_2105_begin_0 = const()[name = tensor("op_2105_begin_0"), val = tensor([0, 2, 0])]; + tensor var_2105_end_0 = const()[name = tensor("op_2105_end_0"), val = tensor([1, 70, 1024])]; + tensor var_2105_end_mask_0 = const()[name = tensor("op_2105_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2105 = slice_by_index(begin = var_2105_begin_0, end = var_2105_end_0, end_mask = var_2105_end_mask_0, x = cache_33)[name = tensor("op_2105")]; + tensor var_2111_interleave_0 = const()[name = tensor("op_2111_interleave_0"), val = tensor(false)]; + tensor var_2111 = concat(axis = var_65, interleave = var_2111_interleave_0, values = (var_2105, key_17))[name = tensor("op_2111")]; + tensor var_2114 = linear(bias = linear_2_bias_0, weight = module_layers_8_self_attn_linear_q_weight_quantized, x = key_17)[name = tensor("linear_75")]; + tensor var_2115 = const()[name = tensor("op_2115"), val = tensor([1, -1, 8, 128])]; + tensor q_49 = reshape(shape = var_2115, x = var_2114)[name = tensor("q_49")]; + tensor var_2118 = linear(bias = linear_2_bias_0, weight = module_layers_8_self_attn_linear_k_weight_quantized, x = input_455)[name = tensor("linear_76")]; + tensor var_2119 = const()[name = tensor("op_2119"), val = tensor([1, -1, 8, 128])]; + tensor k_33 = reshape(shape = var_2119, x = var_2118)[name = tensor("k_33")]; + tensor var_2122 = linear(bias = linear_2_bias_0, weight = module_layers_8_self_attn_linear_v_weight_quantized, x = input_455)[name = tensor("linear_77")]; + tensor var_2123 = const()[name = tensor("op_2123"), val = tensor([1, -1, 8, 128])]; + tensor v_17 = reshape(shape = var_2123, x = var_2122)[name = tensor("v_17")]; + tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_2135 = add(x = q_49, y = module_layers_8_self_attn_pos_bias_u)[name = tensor("op_2135")]; + tensor var_2137 = add(x = q_49, y = module_layers_8_self_attn_pos_bias_v)[name = tensor("op_2137")]; + tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_2139_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2139_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588512320))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588659072))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588658816)))]; + tensor x_215_transpose_x_0 = const()[name = tensor("x_215_transpose_x_0"), val = tensor(false)]; + tensor x_215_transpose_y_0 = const()[name = tensor("x_215_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_17 = transpose(perm = q_with_bias_v_17_perm_0, x = var_2137)[name = tensor("transpose_290")]; + tensor x_215 = matmul(transpose_x = x_215_transpose_x_0, transpose_y = x_215_transpose_y_0, x = q_with_bias_v_17, y = op_2139_quantized)[name = tensor("x_215")]; + tensor const_183 = const()[name = tensor("const_183"), val = tensor(0x0p+0)]; + tensor x_217_pad_0 = const()[name = tensor("x_217_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_217_mode_0 = const()[name = tensor("x_217_mode_0"), val = tensor("constant")]; + tensor x_217 = pad(constant_val = const_183, mode = x_217_mode_0, pad = x_217_pad_0, x = x_215)[name = tensor("x_217")]; + tensor var_2147 = const()[name = tensor("op_2147"), val = tensor([1, 8, -1, 2])]; + tensor x_219 = reshape(shape = var_2147, x = x_217)[name = tensor("x_219")]; + tensor var_2151_begin_0 = const()[name = tensor("op_2151_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2151_end_0 = const()[name = tensor("op_2151_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_2151_end_mask_0 = const()[name = tensor("op_2151_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2151 = slice_by_index(begin = var_2151_begin_0, end = var_2151_end_0, end_mask = var_2151_end_mask_0, x = x_219)[name = tensor("op_2151")]; + tensor var_2152 = const()[name = tensor("op_2152"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_33 = reshape(shape = var_2152, x = var_2151)[name = tensor("matrix_bd_33")]; + tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; + tensor transpose_112_perm_0 = const()[name = tensor("transpose_112_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_113_perm_0 = const()[name = tensor("transpose_113_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_113 = transpose(perm = transpose_113_perm_0, x = k_33)[name = tensor("transpose_288")]; + tensor transpose_112 = transpose(perm = transpose_112_perm_0, x = var_2135)[name = tensor("transpose_289")]; + tensor matrix_ac_17 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_112, y = transpose_113)[name = tensor("matrix_ac_17")]; + tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_35 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33)[name = tensor("matrix_bd_35")]; + tensor var_2161 = add(x = matrix_ac_17, y = matrix_bd_35)[name = tensor("op_2161")]; + tensor _inversed_scores_33_y_0 = const()[name = tensor("_inversed_scores_33_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_33 = mul(x = var_2161, y = _inversed_scores_33_y_0)[name = tensor("_inversed_scores_33")]; + tensor scores_35 = select(a = var_41, b = _inversed_scores_33, cond = mask_11)[name = tensor("scores_35")]; + tensor var_2167 = softmax(axis = var_56, x = scores_35)[name = tensor("op_2167")]; + tensor input_457 = select(a = var_40, b = var_2167, cond = mask_11)[name = tensor("input_457")]; + tensor x_221_transpose_x_0 = const()[name = tensor("x_221_transpose_x_0"), val = tensor(false)]; + tensor x_221_transpose_y_0 = const()[name = tensor("x_221_transpose_y_0"), val = tensor(false)]; + tensor value_25 = transpose(perm = value_25_perm_0, x = v_17)[name = tensor("transpose_287")]; + tensor x_221 = matmul(transpose_x = x_221_transpose_x_0, transpose_y = x_221_transpose_y_0, x = input_457, y = value_25)[name = tensor("x_221")]; + tensor var_2171_perm_0 = const()[name = tensor("op_2171_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2172 = const()[name = tensor("op_2172"), val = tensor([1, -1, 1024])]; + tensor var_2171 = transpose(perm = var_2171_perm_0, x = x_221)[name = tensor("transpose_286")]; + tensor input_459 = reshape(shape = var_2172, x = var_2171)[name = tensor("input_459")]; + tensor input_461 = linear(bias = linear_2_bias_0, weight = module_layers_8_self_attn_linear_out_weight_quantized, x = input_459)[name = tensor("linear_79")]; + tensor input_463 = add(x = input_453, y = input_461)[name = tensor("input_463")]; + tensor x_225_axes_0 = const()[name = tensor("x_225_axes_0"), val = tensor([-1])]; + tensor x_225 = layer_norm(axes = x_225_axes_0, beta = module_layers_8_norm_conv_bias, epsilon = var_38, gamma = module_layers_8_norm_conv_weight, x = input_463)[name = tensor("x_225")]; + tensor input_465_perm_0 = const()[name = tensor("input_465_perm_0"), val = tensor([0, 2, 1])]; + tensor input_467_pad_type_0 = const()[name = tensor("input_467_pad_type_0"), val = tensor("valid")]; + tensor input_467_strides_0 = const()[name = tensor("input_467_strides_0"), val = tensor([1])]; + tensor input_467_pad_0 = const()[name = tensor("input_467_pad_0"), val = tensor([0, 0])]; + tensor input_467_dilations_0 = const()[name = tensor("input_467_dilations_0"), val = tensor([1])]; + tensor input_467_groups_0 = const()[name = tensor("input_467_groups_0"), val = tensor(1)]; + tensor input_465 = transpose(perm = input_465_perm_0, x = x_225)[name = tensor("transpose_285")]; + tensor input_467 = conv(dilations = input_467_dilations_0, groups = input_467_groups_0, pad = input_467_pad_0, pad_type = input_467_pad_type_0, strides = input_467_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_quantized, x = input_465)[name = tensor("input_467")]; + tensor x_227_split_num_splits_0 = const()[name = tensor("x_227_split_num_splits_0"), val = tensor(2)]; + tensor x_227_split_axis_0 = const()[name = tensor("x_227_split_axis_0"), val = tensor(1)]; + tensor x_227_split_0, tensor x_227_split_1 = split(axis = x_227_split_axis_0, num_splits = x_227_split_num_splits_0, x = input_467)[name = tensor("x_227_split")]; + tensor x_227_split_1_sigmoid = sigmoid(x = x_227_split_1)[name = tensor("x_227_split_1_sigmoid")]; + tensor x_227 = mul(x = x_227_split_0, y = x_227_split_1_sigmoid)[name = tensor("x_227")]; + tensor input_469 = select(a = var_40, b = x_227, cond = var_565)[name = tensor("input_469")]; + tensor new_x_35_interleave_0 = const()[name = tensor("new_x_35_interleave_0"), val = tensor(false)]; + tensor new_x_35 = concat(axis = var_56, interleave = new_x_35_interleave_0, values = (cache_35, input_469))[name = tensor("new_x_35")]; + tensor var_2210_begin_0 = const()[name = tensor("op_2210_begin_0"), val = tensor([0, 0, 2])]; + tensor var_2210_end_0 = const()[name = tensor("op_2210_end_0"), val = tensor([1, 1024, 10])]; + tensor var_2210_end_mask_0 = const()[name = tensor("op_2210_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2210 = slice_by_index(begin = var_2210_begin_0, end = var_2210_end_0, end_mask = var_2210_end_mask_0, x = new_x_35)[name = tensor("op_2210")]; + tensor x_229_pad_type_0 = const()[name = tensor("x_229_pad_type_0"), val = tensor("valid")]; + tensor x_229_groups_0 = const()[name = tensor("x_229_groups_0"), val = tensor(1024)]; + tensor x_229_strides_0 = const()[name = tensor("x_229_strides_0"), val = tensor([1])]; + tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0])]; + tensor x_229_dilations_0 = const()[name = tensor("x_229_dilations_0"), val = tensor([1])]; + tensor x_229 = conv(dilations = x_229_dilations_0, groups = x_229_groups_0, pad = x_229_pad_0, pad_type = x_229_pad_type_0, strides = x_229_strides_0, weight = module_layers_8_conv_depthwise_conv_weight_quantized, x = new_x_35)[name = tensor("x_229")]; + tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; + tensor x_231_axes_0 = const()[name = tensor("x_231_axes_0"), val = tensor([-1])]; + tensor input_471 = transpose(perm = input_471_perm_0, x = x_229)[name = tensor("transpose_284")]; + tensor x_231 = layer_norm(axes = x_231_axes_0, beta = module_layers_8_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_8_conv_batch_norm_weight, x = input_471)[name = tensor("x_231")]; + tensor input_473_perm_0 = const()[name = tensor("input_473_perm_0"), val = tensor([0, 2, 1])]; + tensor input_473 = transpose(perm = input_473_perm_0, x = x_231)[name = tensor("transpose_283")]; + tensor input_475 = silu(x = input_473)[name = tensor("input_475")]; + tensor x_233_pad_type_0 = const()[name = tensor("x_233_pad_type_0"), val = tensor("valid")]; + tensor x_233_strides_0 = const()[name = tensor("x_233_strides_0"), val = tensor([1])]; + tensor x_233_pad_0 = const()[name = tensor("x_233_pad_0"), val = tensor([0, 0])]; + tensor x_233_dilations_0 = const()[name = tensor("x_233_dilations_0"), val = tensor([1])]; + tensor x_233_groups_0 = const()[name = tensor("x_233_groups_0"), val = tensor(1)]; + tensor x_233 = conv(dilations = x_233_dilations_0, groups = x_233_groups_0, pad = x_233_pad_0, pad_type = x_233_pad_type_0, strides = x_233_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_quantized, x = input_475)[name = tensor("x_233")]; + tensor input_477_perm_0 = const()[name = tensor("input_477_perm_0"), val = tensor([0, 2, 1])]; + tensor input_477 = transpose(perm = input_477_perm_0, x = x_233)[name = tensor("transpose_282")]; + tensor input_479 = add(x = input_463, y = input_477)[name = tensor("input_479")]; + tensor input_481_axes_0 = const()[name = tensor("input_481_axes_0"), val = tensor([-1])]; + tensor input_481 = layer_norm(axes = input_481_axes_0, beta = module_layers_8_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_8_norm_feed_forward2_weight, x = input_479)[name = tensor("input_481")]; + tensor input_483 = linear(bias = linear_1_bias_0, weight = module_layers_8_feed_forward2_linear1_weight_quantized, x = input_481)[name = tensor("linear_80")]; + tensor input_485 = silu(x = input_483)[name = tensor("input_485")]; + tensor input_489 = linear(bias = linear_2_bias_0, weight = module_layers_8_feed_forward2_linear2_weight_quantized, x = input_485)[name = tensor("linear_81")]; + tensor var_2251 = const()[name = tensor("op_2251"), val = tensor(0x1p-1)]; + tensor var_2252 = mul(x = input_489, y = var_2251)[name = tensor("op_2252")]; + tensor input_491 = add(x = input_479, y = var_2252)[name = tensor("input_491")]; + tensor input_493_axes_0 = const()[name = tensor("input_493_axes_0"), val = tensor([-1])]; + tensor input_493 = layer_norm(axes = input_493_axes_0, beta = module_layers_8_norm_out_bias, epsilon = var_38, gamma = module_layers_8_norm_out_weight, x = input_491)[name = tensor("input_493")]; + tensor cache_37_begin_0 = const()[name = tensor("cache_37_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_37_end_0 = const()[name = tensor("cache_37_end_0"), val = tensor([10, 1, 70, 1024])]; + tensor cache_37_end_mask_0 = const()[name = tensor("cache_37_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_37_squeeze_mask_0 = const()[name = tensor("cache_37_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_37 = slice_by_index(begin = cache_37_begin_0, end = cache_37_end_0, end_mask = cache_37_end_mask_0, squeeze_mask = cache_37_squeeze_mask_0, x = value_3)[name = tensor("cache_37")]; + tensor cache_39_begin_0 = const()[name = tensor("cache_39_begin_0"), val = tensor([9, 0, 0, 0])]; + tensor cache_39_end_0 = const()[name = tensor("cache_39_end_0"), val = tensor([10, 1, 1024, 8])]; + tensor cache_39_end_mask_0 = const()[name = tensor("cache_39_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_39_squeeze_mask_0 = const()[name = tensor("cache_39_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_39 = slice_by_index(begin = cache_39_begin_0, end = cache_39_end_0, end_mask = cache_39_end_mask_0, squeeze_mask = cache_39_squeeze_mask_0, x = value_5)[name = tensor("cache_39")]; + tensor input_495_axes_0 = const()[name = tensor("input_495_axes_0"), val = tensor([-1])]; + tensor input_495 = layer_norm(axes = input_495_axes_0, beta = module_layers_9_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_9_norm_feed_forward1_weight, x = input_493)[name = tensor("input_495")]; + tensor input_497 = linear(bias = linear_1_bias_0, weight = module_layers_9_feed_forward1_linear1_weight_quantized, x = input_495)[name = tensor("linear_82")]; + tensor input_499 = silu(x = input_497)[name = tensor("input_499")]; + tensor input_503 = linear(bias = linear_2_bias_0, weight = module_layers_9_feed_forward1_linear2_weight_quantized, x = input_499)[name = tensor("linear_83")]; + tensor var_2286 = const()[name = tensor("op_2286"), val = tensor(0x1p-1)]; + tensor var_2287 = mul(x = input_503, y = var_2286)[name = tensor("op_2287")]; + tensor input_505 = add(x = input_493, y = var_2287)[name = tensor("input_505")]; + tensor key_19_axes_0 = const()[name = tensor("key_19_axes_0"), val = tensor([-1])]; + tensor key_19 = layer_norm(axes = key_19_axes_0, beta = module_layers_9_norm_self_att_bias, epsilon = var_38, gamma = module_layers_9_norm_self_att_weight, x = input_505)[name = tensor("key_19")]; + tensor input_507_interleave_0 = const()[name = tensor("input_507_interleave_0"), val = tensor(false)]; + tensor input_507 = concat(axis = var_65, interleave = input_507_interleave_0, values = (cache_37, key_19))[name = tensor("input_507")]; + tensor var_2309_begin_0 = const()[name = tensor("op_2309_begin_0"), val = tensor([0, 2, 0])]; + tensor var_2309_end_0 = const()[name = tensor("op_2309_end_0"), val = tensor([1, 70, 1024])]; + tensor var_2309_end_mask_0 = const()[name = tensor("op_2309_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2309 = slice_by_index(begin = var_2309_begin_0, end = var_2309_end_0, end_mask = var_2309_end_mask_0, x = cache_37)[name = tensor("op_2309")]; + tensor var_2315_interleave_0 = const()[name = tensor("op_2315_interleave_0"), val = tensor(false)]; + tensor var_2315 = concat(axis = var_65, interleave = var_2315_interleave_0, values = (var_2309, key_19))[name = tensor("op_2315")]; + tensor var_2318 = linear(bias = linear_2_bias_0, weight = module_layers_9_self_attn_linear_q_weight_quantized, x = key_19)[name = tensor("linear_84")]; + tensor var_2319 = const()[name = tensor("op_2319"), val = tensor([1, -1, 8, 128])]; + tensor q_55 = reshape(shape = var_2319, x = var_2318)[name = tensor("q_55")]; + tensor var_2322 = linear(bias = linear_2_bias_0, weight = module_layers_9_self_attn_linear_k_weight_quantized, x = input_507)[name = tensor("linear_85")]; + tensor var_2323 = const()[name = tensor("op_2323"), val = tensor([1, -1, 8, 128])]; + tensor k_37 = reshape(shape = var_2323, x = var_2322)[name = tensor("k_37")]; + tensor var_2326 = linear(bias = linear_2_bias_0, weight = module_layers_9_self_attn_linear_v_weight_quantized, x = input_507)[name = tensor("linear_86")]; + tensor var_2327 = const()[name = tensor("op_2327"), val = tensor([1, -1, 8, 128])]; + tensor v_19 = reshape(shape = var_2327, x = var_2326)[name = tensor("v_19")]; + tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_2339 = add(x = q_55, y = module_layers_9_self_attn_pos_bias_u)[name = tensor("op_2339")]; + tensor var_2341 = add(x = q_55, y = module_layers_9_self_attn_pos_bias_v)[name = tensor("op_2341")]; + tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_2343_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2343_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588659712))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588806464))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588806208)))]; + tensor x_241_transpose_x_0 = const()[name = tensor("x_241_transpose_x_0"), val = tensor(false)]; + tensor x_241_transpose_y_0 = const()[name = tensor("x_241_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_19 = transpose(perm = q_with_bias_v_19_perm_0, x = var_2341)[name = tensor("transpose_281")]; + tensor x_241 = matmul(transpose_x = x_241_transpose_x_0, transpose_y = x_241_transpose_y_0, x = q_with_bias_v_19, y = op_2343_quantized)[name = tensor("x_241")]; + tensor const_196 = const()[name = tensor("const_196"), val = tensor(0x0p+0)]; + tensor x_243_pad_0 = const()[name = tensor("x_243_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_243_mode_0 = const()[name = tensor("x_243_mode_0"), val = tensor("constant")]; + tensor x_243 = pad(constant_val = const_196, mode = x_243_mode_0, pad = x_243_pad_0, x = x_241)[name = tensor("x_243")]; + tensor var_2351 = const()[name = tensor("op_2351"), val = tensor([1, 8, -1, 2])]; + tensor x_245 = reshape(shape = var_2351, x = x_243)[name = tensor("x_245")]; + tensor var_2355_begin_0 = const()[name = tensor("op_2355_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2355_end_0 = const()[name = tensor("op_2355_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_2355_end_mask_0 = const()[name = tensor("op_2355_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2355 = slice_by_index(begin = var_2355_begin_0, end = var_2355_end_0, end_mask = var_2355_end_mask_0, x = x_245)[name = tensor("op_2355")]; + tensor var_2356 = const()[name = tensor("op_2356"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_37 = reshape(shape = var_2356, x = var_2355)[name = tensor("matrix_bd_37")]; + tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; + tensor transpose_114_perm_0 = const()[name = tensor("transpose_114_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_115_perm_0 = const()[name = tensor("transpose_115_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_115 = transpose(perm = transpose_115_perm_0, x = k_37)[name = tensor("transpose_279")]; + tensor transpose_114 = transpose(perm = transpose_114_perm_0, x = var_2339)[name = tensor("transpose_280")]; + tensor matrix_ac_19 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_114, y = transpose_115)[name = tensor("matrix_ac_19")]; + tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_39 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37)[name = tensor("matrix_bd_39")]; + tensor var_2365 = add(x = matrix_ac_19, y = matrix_bd_39)[name = tensor("op_2365")]; + tensor _inversed_scores_37_y_0 = const()[name = tensor("_inversed_scores_37_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_37 = mul(x = var_2365, y = _inversed_scores_37_y_0)[name = tensor("_inversed_scores_37")]; + tensor scores_39 = select(a = var_41, b = _inversed_scores_37, cond = mask_11)[name = tensor("scores_39")]; + tensor var_2371 = softmax(axis = var_56, x = scores_39)[name = tensor("op_2371")]; + tensor input_509 = select(a = var_40, b = var_2371, cond = mask_11)[name = tensor("input_509")]; + tensor x_247_transpose_x_0 = const()[name = tensor("x_247_transpose_x_0"), val = tensor(false)]; + tensor x_247_transpose_y_0 = const()[name = tensor("x_247_transpose_y_0"), val = tensor(false)]; + tensor value_27 = transpose(perm = value_27_perm_0, x = v_19)[name = tensor("transpose_278")]; + tensor x_247 = matmul(transpose_x = x_247_transpose_x_0, transpose_y = x_247_transpose_y_0, x = input_509, y = value_27)[name = tensor("x_247")]; + tensor var_2375_perm_0 = const()[name = tensor("op_2375_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2376 = const()[name = tensor("op_2376"), val = tensor([1, -1, 1024])]; + tensor var_2375 = transpose(perm = var_2375_perm_0, x = x_247)[name = tensor("transpose_277")]; + tensor input_511 = reshape(shape = var_2376, x = var_2375)[name = tensor("input_511")]; + tensor input_513 = linear(bias = linear_2_bias_0, weight = module_layers_9_self_attn_linear_out_weight_quantized, x = input_511)[name = tensor("linear_88")]; + tensor input_515 = add(x = input_505, y = input_513)[name = tensor("input_515")]; + tensor x_251_axes_0 = const()[name = tensor("x_251_axes_0"), val = tensor([-1])]; + tensor x_251 = layer_norm(axes = x_251_axes_0, beta = module_layers_9_norm_conv_bias, epsilon = var_38, gamma = module_layers_9_norm_conv_weight, x = input_515)[name = tensor("x_251")]; + tensor input_517_perm_0 = const()[name = tensor("input_517_perm_0"), val = tensor([0, 2, 1])]; + tensor input_519_pad_type_0 = const()[name = tensor("input_519_pad_type_0"), val = tensor("valid")]; + tensor input_519_strides_0 = const()[name = tensor("input_519_strides_0"), val = tensor([1])]; + tensor input_519_pad_0 = const()[name = tensor("input_519_pad_0"), val = tensor([0, 0])]; + tensor input_519_dilations_0 = const()[name = tensor("input_519_dilations_0"), val = tensor([1])]; + tensor input_519_groups_0 = const()[name = tensor("input_519_groups_0"), val = tensor(1)]; + tensor input_517 = transpose(perm = input_517_perm_0, x = x_251)[name = tensor("transpose_276")]; + tensor input_519 = conv(dilations = input_519_dilations_0, groups = input_519_groups_0, pad = input_519_pad_0, pad_type = input_519_pad_type_0, strides = input_519_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_quantized, x = input_517)[name = tensor("input_519")]; + tensor x_253_split_num_splits_0 = const()[name = tensor("x_253_split_num_splits_0"), val = tensor(2)]; + tensor x_253_split_axis_0 = const()[name = tensor("x_253_split_axis_0"), val = tensor(1)]; + tensor x_253_split_0, tensor x_253_split_1 = split(axis = x_253_split_axis_0, num_splits = x_253_split_num_splits_0, x = input_519)[name = tensor("x_253_split")]; + tensor x_253_split_1_sigmoid = sigmoid(x = x_253_split_1)[name = tensor("x_253_split_1_sigmoid")]; + tensor x_253 = mul(x = x_253_split_0, y = x_253_split_1_sigmoid)[name = tensor("x_253")]; + tensor input_521 = select(a = var_40, b = x_253, cond = var_565)[name = tensor("input_521")]; + tensor new_x_39_interleave_0 = const()[name = tensor("new_x_39_interleave_0"), val = tensor(false)]; + tensor new_x_39 = concat(axis = var_56, interleave = new_x_39_interleave_0, values = (cache_39, input_521))[name = tensor("new_x_39")]; + tensor var_2414_begin_0 = const()[name = tensor("op_2414_begin_0"), val = tensor([0, 0, 2])]; + tensor var_2414_end_0 = const()[name = tensor("op_2414_end_0"), val = tensor([1, 1024, 10])]; + tensor var_2414_end_mask_0 = const()[name = tensor("op_2414_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2414 = slice_by_index(begin = var_2414_begin_0, end = var_2414_end_0, end_mask = var_2414_end_mask_0, x = new_x_39)[name = tensor("op_2414")]; + tensor x_255_pad_type_0 = const()[name = tensor("x_255_pad_type_0"), val = tensor("valid")]; + tensor x_255_groups_0 = const()[name = tensor("x_255_groups_0"), val = tensor(1024)]; + tensor x_255_strides_0 = const()[name = tensor("x_255_strides_0"), val = tensor([1])]; + tensor x_255_pad_0 = const()[name = tensor("x_255_pad_0"), val = tensor([0, 0])]; + tensor x_255_dilations_0 = const()[name = tensor("x_255_dilations_0"), val = tensor([1])]; + tensor x_255 = conv(dilations = x_255_dilations_0, groups = x_255_groups_0, pad = x_255_pad_0, pad_type = x_255_pad_type_0, strides = x_255_strides_0, weight = module_layers_9_conv_depthwise_conv_weight_quantized, x = new_x_39)[name = tensor("x_255")]; + tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; + tensor x_257_axes_0 = const()[name = tensor("x_257_axes_0"), val = tensor([-1])]; + tensor input_523 = transpose(perm = input_523_perm_0, x = x_255)[name = tensor("transpose_275")]; + tensor x_257 = layer_norm(axes = x_257_axes_0, beta = module_layers_9_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_9_conv_batch_norm_weight, x = input_523)[name = tensor("x_257")]; + tensor input_525_perm_0 = const()[name = tensor("input_525_perm_0"), val = tensor([0, 2, 1])]; + tensor input_525 = transpose(perm = input_525_perm_0, x = x_257)[name = tensor("transpose_274")]; + tensor input_527 = silu(x = input_525)[name = tensor("input_527")]; + tensor x_259_pad_type_0 = const()[name = tensor("x_259_pad_type_0"), val = tensor("valid")]; + tensor x_259_strides_0 = const()[name = tensor("x_259_strides_0"), val = tensor([1])]; + tensor x_259_pad_0 = const()[name = tensor("x_259_pad_0"), val = tensor([0, 0])]; + tensor x_259_dilations_0 = const()[name = tensor("x_259_dilations_0"), val = tensor([1])]; + tensor x_259_groups_0 = const()[name = tensor("x_259_groups_0"), val = tensor(1)]; + tensor x_259 = conv(dilations = x_259_dilations_0, groups = x_259_groups_0, pad = x_259_pad_0, pad_type = x_259_pad_type_0, strides = x_259_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_quantized, x = input_527)[name = tensor("x_259")]; + tensor input_529_perm_0 = const()[name = tensor("input_529_perm_0"), val = tensor([0, 2, 1])]; + tensor input_529 = transpose(perm = input_529_perm_0, x = x_259)[name = tensor("transpose_273")]; + tensor input_531 = add(x = input_515, y = input_529)[name = tensor("input_531")]; + tensor input_533_axes_0 = const()[name = tensor("input_533_axes_0"), val = tensor([-1])]; + tensor input_533 = layer_norm(axes = input_533_axes_0, beta = module_layers_9_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_9_norm_feed_forward2_weight, x = input_531)[name = tensor("input_533")]; + tensor input_535 = linear(bias = linear_1_bias_0, weight = module_layers_9_feed_forward2_linear1_weight_quantized, x = input_533)[name = tensor("linear_89")]; + tensor input_537 = silu(x = input_535)[name = tensor("input_537")]; + tensor input_541 = linear(bias = linear_2_bias_0, weight = module_layers_9_feed_forward2_linear2_weight_quantized, x = input_537)[name = tensor("linear_90")]; + tensor var_2455 = const()[name = tensor("op_2455"), val = tensor(0x1p-1)]; + tensor var_2456 = mul(x = input_541, y = var_2455)[name = tensor("op_2456")]; + tensor input_543 = add(x = input_531, y = var_2456)[name = tensor("input_543")]; + tensor input_545_axes_0 = const()[name = tensor("input_545_axes_0"), val = tensor([-1])]; + tensor input_545 = layer_norm(axes = input_545_axes_0, beta = module_layers_9_norm_out_bias, epsilon = var_38, gamma = module_layers_9_norm_out_weight, x = input_543)[name = tensor("input_545")]; + tensor cache_41_begin_0 = const()[name = tensor("cache_41_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_41_end_0 = const()[name = tensor("cache_41_end_0"), val = tensor([11, 1, 70, 1024])]; + tensor cache_41_end_mask_0 = const()[name = tensor("cache_41_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_41_squeeze_mask_0 = const()[name = tensor("cache_41_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_41 = slice_by_index(begin = cache_41_begin_0, end = cache_41_end_0, end_mask = cache_41_end_mask_0, squeeze_mask = cache_41_squeeze_mask_0, x = value_3)[name = tensor("cache_41")]; + tensor cache_43_begin_0 = const()[name = tensor("cache_43_begin_0"), val = tensor([10, 0, 0, 0])]; + tensor cache_43_end_0 = const()[name = tensor("cache_43_end_0"), val = tensor([11, 1, 1024, 8])]; + tensor cache_43_end_mask_0 = const()[name = tensor("cache_43_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_43_squeeze_mask_0 = const()[name = tensor("cache_43_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_43 = slice_by_index(begin = cache_43_begin_0, end = cache_43_end_0, end_mask = cache_43_end_mask_0, squeeze_mask = cache_43_squeeze_mask_0, x = value_5)[name = tensor("cache_43")]; + tensor input_547_axes_0 = const()[name = tensor("input_547_axes_0"), val = tensor([-1])]; + tensor input_547 = layer_norm(axes = input_547_axes_0, beta = module_layers_10_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_10_norm_feed_forward1_weight, x = input_545)[name = tensor("input_547")]; + tensor input_549 = linear(bias = linear_1_bias_0, weight = module_layers_10_feed_forward1_linear1_weight_quantized, x = input_547)[name = tensor("linear_91")]; + tensor input_551 = silu(x = input_549)[name = tensor("input_551")]; + tensor input_555 = linear(bias = linear_2_bias_0, weight = module_layers_10_feed_forward1_linear2_weight_quantized, x = input_551)[name = tensor("linear_92")]; + tensor var_2490 = const()[name = tensor("op_2490"), val = tensor(0x1p-1)]; + tensor var_2491 = mul(x = input_555, y = var_2490)[name = tensor("op_2491")]; + tensor input_557 = add(x = input_545, y = var_2491)[name = tensor("input_557")]; + tensor key_21_axes_0 = const()[name = tensor("key_21_axes_0"), val = tensor([-1])]; + tensor key_21 = layer_norm(axes = key_21_axes_0, beta = module_layers_10_norm_self_att_bias, epsilon = var_38, gamma = module_layers_10_norm_self_att_weight, x = input_557)[name = tensor("key_21")]; + tensor input_559_interleave_0 = const()[name = tensor("input_559_interleave_0"), val = tensor(false)]; + tensor input_559 = concat(axis = var_65, interleave = input_559_interleave_0, values = (cache_41, key_21))[name = tensor("input_559")]; + tensor var_2513_begin_0 = const()[name = tensor("op_2513_begin_0"), val = tensor([0, 2, 0])]; + tensor var_2513_end_0 = const()[name = tensor("op_2513_end_0"), val = tensor([1, 70, 1024])]; + tensor var_2513_end_mask_0 = const()[name = tensor("op_2513_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2513 = slice_by_index(begin = var_2513_begin_0, end = var_2513_end_0, end_mask = var_2513_end_mask_0, x = cache_41)[name = tensor("op_2513")]; + tensor var_2519_interleave_0 = const()[name = tensor("op_2519_interleave_0"), val = tensor(false)]; + tensor var_2519 = concat(axis = var_65, interleave = var_2519_interleave_0, values = (var_2513, key_21))[name = tensor("op_2519")]; + tensor var_2522 = linear(bias = linear_2_bias_0, weight = module_layers_10_self_attn_linear_q_weight_quantized, x = key_21)[name = tensor("linear_93")]; + tensor var_2523 = const()[name = tensor("op_2523"), val = tensor([1, -1, 8, 128])]; + tensor q_61 = reshape(shape = var_2523, x = var_2522)[name = tensor("q_61")]; + tensor var_2526 = linear(bias = linear_2_bias_0, weight = module_layers_10_self_attn_linear_k_weight_quantized, x = input_559)[name = tensor("linear_94")]; + tensor var_2527 = const()[name = tensor("op_2527"), val = tensor([1, -1, 8, 128])]; + tensor k_41 = reshape(shape = var_2527, x = var_2526)[name = tensor("k_41")]; + tensor var_2530 = linear(bias = linear_2_bias_0, weight = module_layers_10_self_attn_linear_v_weight_quantized, x = input_559)[name = tensor("linear_95")]; + tensor var_2531 = const()[name = tensor("op_2531"), val = tensor([1, -1, 8, 128])]; + tensor v_21 = reshape(shape = var_2531, x = var_2530)[name = tensor("v_21")]; + tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_2543 = add(x = q_61, y = module_layers_10_self_attn_pos_bias_u)[name = tensor("op_2543")]; + tensor var_2545 = add(x = q_61, y = module_layers_10_self_attn_pos_bias_v)[name = tensor("op_2545")]; + tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_2547_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2547_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588807104))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588953856))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588953600)))]; + tensor x_267_transpose_x_0 = const()[name = tensor("x_267_transpose_x_0"), val = tensor(false)]; + tensor x_267_transpose_y_0 = const()[name = tensor("x_267_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_21 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2545)[name = tensor("transpose_272")]; + tensor x_267 = matmul(transpose_x = x_267_transpose_x_0, transpose_y = x_267_transpose_y_0, x = q_with_bias_v_21, y = op_2547_quantized)[name = tensor("x_267")]; + tensor const_209 = const()[name = tensor("const_209"), val = tensor(0x0p+0)]; + tensor x_269_pad_0 = const()[name = tensor("x_269_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_269_mode_0 = const()[name = tensor("x_269_mode_0"), val = tensor("constant")]; + tensor x_269 = pad(constant_val = const_209, mode = x_269_mode_0, pad = x_269_pad_0, x = x_267)[name = tensor("x_269")]; + tensor var_2555 = const()[name = tensor("op_2555"), val = tensor([1, 8, -1, 2])]; + tensor x_271 = reshape(shape = var_2555, x = x_269)[name = tensor("x_271")]; + tensor var_2559_begin_0 = const()[name = tensor("op_2559_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2559_end_0 = const()[name = tensor("op_2559_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_2559_end_mask_0 = const()[name = tensor("op_2559_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2559 = slice_by_index(begin = var_2559_begin_0, end = var_2559_end_0, end_mask = var_2559_end_mask_0, x = x_271)[name = tensor("op_2559")]; + tensor var_2560 = const()[name = tensor("op_2560"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_41 = reshape(shape = var_2560, x = var_2559)[name = tensor("matrix_bd_41")]; + tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; + tensor transpose_116_perm_0 = const()[name = tensor("transpose_116_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_117_perm_0 = const()[name = tensor("transpose_117_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_117 = transpose(perm = transpose_117_perm_0, x = k_41)[name = tensor("transpose_270")]; + tensor transpose_116 = transpose(perm = transpose_116_perm_0, x = var_2543)[name = tensor("transpose_271")]; + tensor matrix_ac_21 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_116, y = transpose_117)[name = tensor("matrix_ac_21")]; + tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_43 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41)[name = tensor("matrix_bd_43")]; + tensor var_2569 = add(x = matrix_ac_21, y = matrix_bd_43)[name = tensor("op_2569")]; + tensor _inversed_scores_41_y_0 = const()[name = tensor("_inversed_scores_41_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_41 = mul(x = var_2569, y = _inversed_scores_41_y_0)[name = tensor("_inversed_scores_41")]; + tensor scores_43 = select(a = var_41, b = _inversed_scores_41, cond = mask_11)[name = tensor("scores_43")]; + tensor var_2575 = softmax(axis = var_56, x = scores_43)[name = tensor("op_2575")]; + tensor input_561 = select(a = var_40, b = var_2575, cond = mask_11)[name = tensor("input_561")]; + tensor x_273_transpose_x_0 = const()[name = tensor("x_273_transpose_x_0"), val = tensor(false)]; + tensor x_273_transpose_y_0 = const()[name = tensor("x_273_transpose_y_0"), val = tensor(false)]; + tensor value_29 = transpose(perm = value_29_perm_0, x = v_21)[name = tensor("transpose_269")]; + tensor x_273 = matmul(transpose_x = x_273_transpose_x_0, transpose_y = x_273_transpose_y_0, x = input_561, y = value_29)[name = tensor("x_273")]; + tensor var_2579_perm_0 = const()[name = tensor("op_2579_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2580 = const()[name = tensor("op_2580"), val = tensor([1, -1, 1024])]; + tensor var_2579 = transpose(perm = var_2579_perm_0, x = x_273)[name = tensor("transpose_268")]; + tensor input_563 = reshape(shape = var_2580, x = var_2579)[name = tensor("input_563")]; + tensor input_565 = linear(bias = linear_2_bias_0, weight = module_layers_10_self_attn_linear_out_weight_quantized, x = input_563)[name = tensor("linear_97")]; + tensor input_567 = add(x = input_557, y = input_565)[name = tensor("input_567")]; + tensor x_277_axes_0 = const()[name = tensor("x_277_axes_0"), val = tensor([-1])]; + tensor x_277 = layer_norm(axes = x_277_axes_0, beta = module_layers_10_norm_conv_bias, epsilon = var_38, gamma = module_layers_10_norm_conv_weight, x = input_567)[name = tensor("x_277")]; + tensor input_569_perm_0 = const()[name = tensor("input_569_perm_0"), val = tensor([0, 2, 1])]; + tensor input_571_pad_type_0 = const()[name = tensor("input_571_pad_type_0"), val = tensor("valid")]; + tensor input_571_strides_0 = const()[name = tensor("input_571_strides_0"), val = tensor([1])]; + tensor input_571_pad_0 = const()[name = tensor("input_571_pad_0"), val = tensor([0, 0])]; + tensor input_571_dilations_0 = const()[name = tensor("input_571_dilations_0"), val = tensor([1])]; + tensor input_571_groups_0 = const()[name = tensor("input_571_groups_0"), val = tensor(1)]; + tensor input_569 = transpose(perm = input_569_perm_0, x = x_277)[name = tensor("transpose_267")]; + tensor input_571 = conv(dilations = input_571_dilations_0, groups = input_571_groups_0, pad = input_571_pad_0, pad_type = input_571_pad_type_0, strides = input_571_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_quantized, x = input_569)[name = tensor("input_571")]; + tensor x_279_split_num_splits_0 = const()[name = tensor("x_279_split_num_splits_0"), val = tensor(2)]; + tensor x_279_split_axis_0 = const()[name = tensor("x_279_split_axis_0"), val = tensor(1)]; + tensor x_279_split_0, tensor x_279_split_1 = split(axis = x_279_split_axis_0, num_splits = x_279_split_num_splits_0, x = input_571)[name = tensor("x_279_split")]; + tensor x_279_split_1_sigmoid = sigmoid(x = x_279_split_1)[name = tensor("x_279_split_1_sigmoid")]; + tensor x_279 = mul(x = x_279_split_0, y = x_279_split_1_sigmoid)[name = tensor("x_279")]; + tensor input_573 = select(a = var_40, b = x_279, cond = var_565)[name = tensor("input_573")]; + tensor new_x_43_interleave_0 = const()[name = tensor("new_x_43_interleave_0"), val = tensor(false)]; + tensor new_x_43 = concat(axis = var_56, interleave = new_x_43_interleave_0, values = (cache_43, input_573))[name = tensor("new_x_43")]; + tensor var_2618_begin_0 = const()[name = tensor("op_2618_begin_0"), val = tensor([0, 0, 2])]; + tensor var_2618_end_0 = const()[name = tensor("op_2618_end_0"), val = tensor([1, 1024, 10])]; + tensor var_2618_end_mask_0 = const()[name = tensor("op_2618_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2618 = slice_by_index(begin = var_2618_begin_0, end = var_2618_end_0, end_mask = var_2618_end_mask_0, x = new_x_43)[name = tensor("op_2618")]; + tensor x_281_pad_type_0 = const()[name = tensor("x_281_pad_type_0"), val = tensor("valid")]; + tensor x_281_groups_0 = const()[name = tensor("x_281_groups_0"), val = tensor(1024)]; + tensor x_281_strides_0 = const()[name = tensor("x_281_strides_0"), val = tensor([1])]; + tensor x_281_pad_0 = const()[name = tensor("x_281_pad_0"), val = tensor([0, 0])]; + tensor x_281_dilations_0 = const()[name = tensor("x_281_dilations_0"), val = tensor([1])]; + tensor x_281 = conv(dilations = x_281_dilations_0, groups = x_281_groups_0, pad = x_281_pad_0, pad_type = x_281_pad_type_0, strides = x_281_strides_0, weight = module_layers_10_conv_depthwise_conv_weight_quantized, x = new_x_43)[name = tensor("x_281")]; + tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; + tensor x_283_axes_0 = const()[name = tensor("x_283_axes_0"), val = tensor([-1])]; + tensor input_575 = transpose(perm = input_575_perm_0, x = x_281)[name = tensor("transpose_266")]; + tensor x_283 = layer_norm(axes = x_283_axes_0, beta = module_layers_10_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_10_conv_batch_norm_weight, x = input_575)[name = tensor("x_283")]; + tensor input_577_perm_0 = const()[name = tensor("input_577_perm_0"), val = tensor([0, 2, 1])]; + tensor input_577 = transpose(perm = input_577_perm_0, x = x_283)[name = tensor("transpose_265")]; + tensor input_579 = silu(x = input_577)[name = tensor("input_579")]; + tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; + tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; + tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; + tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; + tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; + tensor x_285 = conv(dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_quantized, x = input_579)[name = tensor("x_285")]; + tensor input_581_perm_0 = const()[name = tensor("input_581_perm_0"), val = tensor([0, 2, 1])]; + tensor input_581 = transpose(perm = input_581_perm_0, x = x_285)[name = tensor("transpose_264")]; + tensor input_583 = add(x = input_567, y = input_581)[name = tensor("input_583")]; + tensor input_585_axes_0 = const()[name = tensor("input_585_axes_0"), val = tensor([-1])]; + tensor input_585 = layer_norm(axes = input_585_axes_0, beta = module_layers_10_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_10_norm_feed_forward2_weight, x = input_583)[name = tensor("input_585")]; + tensor input_587 = linear(bias = linear_1_bias_0, weight = module_layers_10_feed_forward2_linear1_weight_quantized, x = input_585)[name = tensor("linear_98")]; + tensor input_589 = silu(x = input_587)[name = tensor("input_589")]; + tensor input_593 = linear(bias = linear_2_bias_0, weight = module_layers_10_feed_forward2_linear2_weight_quantized, x = input_589)[name = tensor("linear_99")]; + tensor var_2659 = const()[name = tensor("op_2659"), val = tensor(0x1p-1)]; + tensor var_2660 = mul(x = input_593, y = var_2659)[name = tensor("op_2660")]; + tensor input_595 = add(x = input_583, y = var_2660)[name = tensor("input_595")]; + tensor input_597_axes_0 = const()[name = tensor("input_597_axes_0"), val = tensor([-1])]; + tensor input_597 = layer_norm(axes = input_597_axes_0, beta = module_layers_10_norm_out_bias, epsilon = var_38, gamma = module_layers_10_norm_out_weight, x = input_595)[name = tensor("input_597")]; + tensor cache_45_begin_0 = const()[name = tensor("cache_45_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_45_end_0 = const()[name = tensor("cache_45_end_0"), val = tensor([12, 1, 70, 1024])]; + tensor cache_45_end_mask_0 = const()[name = tensor("cache_45_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_45_squeeze_mask_0 = const()[name = tensor("cache_45_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_45 = slice_by_index(begin = cache_45_begin_0, end = cache_45_end_0, end_mask = cache_45_end_mask_0, squeeze_mask = cache_45_squeeze_mask_0, x = value_3)[name = tensor("cache_45")]; + tensor cache_47_begin_0 = const()[name = tensor("cache_47_begin_0"), val = tensor([11, 0, 0, 0])]; + tensor cache_47_end_0 = const()[name = tensor("cache_47_end_0"), val = tensor([12, 1, 1024, 8])]; + tensor cache_47_end_mask_0 = const()[name = tensor("cache_47_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_47_squeeze_mask_0 = const()[name = tensor("cache_47_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_47 = slice_by_index(begin = cache_47_begin_0, end = cache_47_end_0, end_mask = cache_47_end_mask_0, squeeze_mask = cache_47_squeeze_mask_0, x = value_5)[name = tensor("cache_47")]; + tensor input_599_axes_0 = const()[name = tensor("input_599_axes_0"), val = tensor([-1])]; + tensor input_599 = layer_norm(axes = input_599_axes_0, beta = module_layers_11_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_11_norm_feed_forward1_weight, x = input_597)[name = tensor("input_599")]; + tensor input_601 = linear(bias = linear_1_bias_0, weight = module_layers_11_feed_forward1_linear1_weight_quantized, x = input_599)[name = tensor("linear_100")]; + tensor input_603 = silu(x = input_601)[name = tensor("input_603")]; + tensor input_607 = linear(bias = linear_2_bias_0, weight = module_layers_11_feed_forward1_linear2_weight_quantized, x = input_603)[name = tensor("linear_101")]; + tensor var_2694 = const()[name = tensor("op_2694"), val = tensor(0x1p-1)]; + tensor var_2695 = mul(x = input_607, y = var_2694)[name = tensor("op_2695")]; + tensor input_609 = add(x = input_597, y = var_2695)[name = tensor("input_609")]; + tensor key_23_axes_0 = const()[name = tensor("key_23_axes_0"), val = tensor([-1])]; + tensor key_23 = layer_norm(axes = key_23_axes_0, beta = module_layers_11_norm_self_att_bias, epsilon = var_38, gamma = module_layers_11_norm_self_att_weight, x = input_609)[name = tensor("key_23")]; + tensor input_611_interleave_0 = const()[name = tensor("input_611_interleave_0"), val = tensor(false)]; + tensor input_611 = concat(axis = var_65, interleave = input_611_interleave_0, values = (cache_45, key_23))[name = tensor("input_611")]; + tensor var_2717_begin_0 = const()[name = tensor("op_2717_begin_0"), val = tensor([0, 2, 0])]; + tensor var_2717_end_0 = const()[name = tensor("op_2717_end_0"), val = tensor([1, 70, 1024])]; + tensor var_2717_end_mask_0 = const()[name = tensor("op_2717_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2717 = slice_by_index(begin = var_2717_begin_0, end = var_2717_end_0, end_mask = var_2717_end_mask_0, x = cache_45)[name = tensor("op_2717")]; + tensor var_2723_interleave_0 = const()[name = tensor("op_2723_interleave_0"), val = tensor(false)]; + tensor var_2723 = concat(axis = var_65, interleave = var_2723_interleave_0, values = (var_2717, key_23))[name = tensor("op_2723")]; + tensor var_2726 = linear(bias = linear_2_bias_0, weight = module_layers_11_self_attn_linear_q_weight_quantized, x = key_23)[name = tensor("linear_102")]; + tensor var_2727 = const()[name = tensor("op_2727"), val = tensor([1, -1, 8, 128])]; + tensor q_67 = reshape(shape = var_2727, x = var_2726)[name = tensor("q_67")]; + tensor var_2730 = linear(bias = linear_2_bias_0, weight = module_layers_11_self_attn_linear_k_weight_quantized, x = input_611)[name = tensor("linear_103")]; + tensor var_2731 = const()[name = tensor("op_2731"), val = tensor([1, -1, 8, 128])]; + tensor k_45 = reshape(shape = var_2731, x = var_2730)[name = tensor("k_45")]; + tensor var_2734 = linear(bias = linear_2_bias_0, weight = module_layers_11_self_attn_linear_v_weight_quantized, x = input_611)[name = tensor("linear_104")]; + tensor var_2735 = const()[name = tensor("op_2735"), val = tensor([1, -1, 8, 128])]; + tensor v_23 = reshape(shape = var_2735, x = var_2734)[name = tensor("v_23")]; + tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_2747 = add(x = q_67, y = module_layers_11_self_attn_pos_bias_u)[name = tensor("op_2747")]; + tensor var_2749 = add(x = q_67, y = module_layers_11_self_attn_pos_bias_v)[name = tensor("op_2749")]; + tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_2751_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2751_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(588954496))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589101248))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589100992)))]; + tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; + tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_23 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2749)[name = tensor("transpose_263")]; + tensor x_293 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_23, y = op_2751_quantized)[name = tensor("x_293")]; + tensor const_222 = const()[name = tensor("const_222"), val = tensor(0x0p+0)]; + tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; + tensor x_295 = pad(constant_val = const_222, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293)[name = tensor("x_295")]; + tensor var_2759 = const()[name = tensor("op_2759"), val = tensor([1, 8, -1, 2])]; + tensor x_297 = reshape(shape = var_2759, x = x_295)[name = tensor("x_297")]; + tensor var_2763_begin_0 = const()[name = tensor("op_2763_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2763_end_0 = const()[name = tensor("op_2763_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_2763_end_mask_0 = const()[name = tensor("op_2763_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2763 = slice_by_index(begin = var_2763_begin_0, end = var_2763_end_0, end_mask = var_2763_end_mask_0, x = x_297)[name = tensor("op_2763")]; + tensor var_2764 = const()[name = tensor("op_2764"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_45 = reshape(shape = var_2764, x = var_2763)[name = tensor("matrix_bd_45")]; + tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; + tensor transpose_118_perm_0 = const()[name = tensor("transpose_118_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_119_perm_0 = const()[name = tensor("transpose_119_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_119 = transpose(perm = transpose_119_perm_0, x = k_45)[name = tensor("transpose_261")]; + tensor transpose_118 = transpose(perm = transpose_118_perm_0, x = var_2747)[name = tensor("transpose_262")]; + tensor matrix_ac_23 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_118, y = transpose_119)[name = tensor("matrix_ac_23")]; + tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_47 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45)[name = tensor("matrix_bd_47")]; + tensor var_2773 = add(x = matrix_ac_23, y = matrix_bd_47)[name = tensor("op_2773")]; + tensor _inversed_scores_45_y_0 = const()[name = tensor("_inversed_scores_45_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_45 = mul(x = var_2773, y = _inversed_scores_45_y_0)[name = tensor("_inversed_scores_45")]; + tensor scores_47 = select(a = var_41, b = _inversed_scores_45, cond = mask_11)[name = tensor("scores_47")]; + tensor var_2779 = softmax(axis = var_56, x = scores_47)[name = tensor("op_2779")]; + tensor input_613 = select(a = var_40, b = var_2779, cond = mask_11)[name = tensor("input_613")]; + tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; + tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; + tensor value_31 = transpose(perm = value_31_perm_0, x = v_23)[name = tensor("transpose_260")]; + tensor x_299 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_613, y = value_31)[name = tensor("x_299")]; + tensor var_2783_perm_0 = const()[name = tensor("op_2783_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2784 = const()[name = tensor("op_2784"), val = tensor([1, -1, 1024])]; + tensor var_2783 = transpose(perm = var_2783_perm_0, x = x_299)[name = tensor("transpose_259")]; + tensor input_615 = reshape(shape = var_2784, x = var_2783)[name = tensor("input_615")]; + tensor input_617 = linear(bias = linear_2_bias_0, weight = module_layers_11_self_attn_linear_out_weight_quantized, x = input_615)[name = tensor("linear_106")]; + tensor input_619 = add(x = input_609, y = input_617)[name = tensor("input_619")]; + tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; + tensor x_303 = layer_norm(axes = x_303_axes_0, beta = module_layers_11_norm_conv_bias, epsilon = var_38, gamma = module_layers_11_norm_conv_weight, x = input_619)[name = tensor("x_303")]; + tensor input_621_perm_0 = const()[name = tensor("input_621_perm_0"), val = tensor([0, 2, 1])]; + tensor input_623_pad_type_0 = const()[name = tensor("input_623_pad_type_0"), val = tensor("valid")]; + tensor input_623_strides_0 = const()[name = tensor("input_623_strides_0"), val = tensor([1])]; + tensor input_623_pad_0 = const()[name = tensor("input_623_pad_0"), val = tensor([0, 0])]; + tensor input_623_dilations_0 = const()[name = tensor("input_623_dilations_0"), val = tensor([1])]; + tensor input_623_groups_0 = const()[name = tensor("input_623_groups_0"), val = tensor(1)]; + tensor input_621 = transpose(perm = input_621_perm_0, x = x_303)[name = tensor("transpose_258")]; + tensor input_623 = conv(dilations = input_623_dilations_0, groups = input_623_groups_0, pad = input_623_pad_0, pad_type = input_623_pad_type_0, strides = input_623_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_quantized, x = input_621)[name = tensor("input_623")]; + tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; + tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; + tensor x_305_split_0, tensor x_305_split_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_623)[name = tensor("x_305_split")]; + tensor x_305_split_1_sigmoid = sigmoid(x = x_305_split_1)[name = tensor("x_305_split_1_sigmoid")]; + tensor x_305 = mul(x = x_305_split_0, y = x_305_split_1_sigmoid)[name = tensor("x_305")]; + tensor input_625 = select(a = var_40, b = x_305, cond = var_565)[name = tensor("input_625")]; + tensor new_x_47_interleave_0 = const()[name = tensor("new_x_47_interleave_0"), val = tensor(false)]; + tensor new_x_47 = concat(axis = var_56, interleave = new_x_47_interleave_0, values = (cache_47, input_625))[name = tensor("new_x_47")]; + tensor var_2822_begin_0 = const()[name = tensor("op_2822_begin_0"), val = tensor([0, 0, 2])]; + tensor var_2822_end_0 = const()[name = tensor("op_2822_end_0"), val = tensor([1, 1024, 10])]; + tensor var_2822_end_mask_0 = const()[name = tensor("op_2822_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2822 = slice_by_index(begin = var_2822_begin_0, end = var_2822_end_0, end_mask = var_2822_end_mask_0, x = new_x_47)[name = tensor("op_2822")]; + tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; + tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1024)]; + tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; + tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; + tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; + tensor x_307 = conv(dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_11_conv_depthwise_conv_weight_quantized, x = new_x_47)[name = tensor("x_307")]; + tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; + tensor x_309_axes_0 = const()[name = tensor("x_309_axes_0"), val = tensor([-1])]; + tensor input_627 = transpose(perm = input_627_perm_0, x = x_307)[name = tensor("transpose_257")]; + tensor x_309 = layer_norm(axes = x_309_axes_0, beta = module_layers_11_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_11_conv_batch_norm_weight, x = input_627)[name = tensor("x_309")]; + tensor input_629_perm_0 = const()[name = tensor("input_629_perm_0"), val = tensor([0, 2, 1])]; + tensor input_629 = transpose(perm = input_629_perm_0, x = x_309)[name = tensor("transpose_256")]; + tensor input_631 = silu(x = input_629)[name = tensor("input_631")]; + tensor x_311_pad_type_0 = const()[name = tensor("x_311_pad_type_0"), val = tensor("valid")]; + tensor x_311_strides_0 = const()[name = tensor("x_311_strides_0"), val = tensor([1])]; + tensor x_311_pad_0 = const()[name = tensor("x_311_pad_0"), val = tensor([0, 0])]; + tensor x_311_dilations_0 = const()[name = tensor("x_311_dilations_0"), val = tensor([1])]; + tensor x_311_groups_0 = const()[name = tensor("x_311_groups_0"), val = tensor(1)]; + tensor x_311 = conv(dilations = x_311_dilations_0, groups = x_311_groups_0, pad = x_311_pad_0, pad_type = x_311_pad_type_0, strides = x_311_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_quantized, x = input_631)[name = tensor("x_311")]; + tensor input_633_perm_0 = const()[name = tensor("input_633_perm_0"), val = tensor([0, 2, 1])]; + tensor input_633 = transpose(perm = input_633_perm_0, x = x_311)[name = tensor("transpose_255")]; + tensor input_635 = add(x = input_619, y = input_633)[name = tensor("input_635")]; + tensor input_637_axes_0 = const()[name = tensor("input_637_axes_0"), val = tensor([-1])]; + tensor input_637 = layer_norm(axes = input_637_axes_0, beta = module_layers_11_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_11_norm_feed_forward2_weight, x = input_635)[name = tensor("input_637")]; + tensor input_639 = linear(bias = linear_1_bias_0, weight = module_layers_11_feed_forward2_linear1_weight_quantized, x = input_637)[name = tensor("linear_107")]; + tensor input_641 = silu(x = input_639)[name = tensor("input_641")]; + tensor input_645 = linear(bias = linear_2_bias_0, weight = module_layers_11_feed_forward2_linear2_weight_quantized, x = input_641)[name = tensor("linear_108")]; + tensor var_2863 = const()[name = tensor("op_2863"), val = tensor(0x1p-1)]; + tensor var_2864 = mul(x = input_645, y = var_2863)[name = tensor("op_2864")]; + tensor input_647 = add(x = input_635, y = var_2864)[name = tensor("input_647")]; + tensor input_649_axes_0 = const()[name = tensor("input_649_axes_0"), val = tensor([-1])]; + tensor input_649 = layer_norm(axes = input_649_axes_0, beta = module_layers_11_norm_out_bias, epsilon = var_38, gamma = module_layers_11_norm_out_weight, x = input_647)[name = tensor("input_649")]; + tensor cache_49_begin_0 = const()[name = tensor("cache_49_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_49_end_0 = const()[name = tensor("cache_49_end_0"), val = tensor([13, 1, 70, 1024])]; + tensor cache_49_end_mask_0 = const()[name = tensor("cache_49_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_49_squeeze_mask_0 = const()[name = tensor("cache_49_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_49 = slice_by_index(begin = cache_49_begin_0, end = cache_49_end_0, end_mask = cache_49_end_mask_0, squeeze_mask = cache_49_squeeze_mask_0, x = value_3)[name = tensor("cache_49")]; + tensor cache_51_begin_0 = const()[name = tensor("cache_51_begin_0"), val = tensor([12, 0, 0, 0])]; + tensor cache_51_end_0 = const()[name = tensor("cache_51_end_0"), val = tensor([13, 1, 1024, 8])]; + tensor cache_51_end_mask_0 = const()[name = tensor("cache_51_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_51_squeeze_mask_0 = const()[name = tensor("cache_51_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_51 = slice_by_index(begin = cache_51_begin_0, end = cache_51_end_0, end_mask = cache_51_end_mask_0, squeeze_mask = cache_51_squeeze_mask_0, x = value_5)[name = tensor("cache_51")]; + tensor input_651_axes_0 = const()[name = tensor("input_651_axes_0"), val = tensor([-1])]; + tensor input_651 = layer_norm(axes = input_651_axes_0, beta = module_layers_12_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_12_norm_feed_forward1_weight, x = input_649)[name = tensor("input_651")]; + tensor input_653 = linear(bias = linear_1_bias_0, weight = module_layers_12_feed_forward1_linear1_weight_quantized, x = input_651)[name = tensor("linear_109")]; + tensor input_655 = silu(x = input_653)[name = tensor("input_655")]; + tensor input_659 = linear(bias = linear_2_bias_0, weight = module_layers_12_feed_forward1_linear2_weight_quantized, x = input_655)[name = tensor("linear_110")]; + tensor var_2898 = const()[name = tensor("op_2898"), val = tensor(0x1p-1)]; + tensor var_2899 = mul(x = input_659, y = var_2898)[name = tensor("op_2899")]; + tensor input_661 = add(x = input_649, y = var_2899)[name = tensor("input_661")]; + tensor key_25_axes_0 = const()[name = tensor("key_25_axes_0"), val = tensor([-1])]; + tensor key_25 = layer_norm(axes = key_25_axes_0, beta = module_layers_12_norm_self_att_bias, epsilon = var_38, gamma = module_layers_12_norm_self_att_weight, x = input_661)[name = tensor("key_25")]; + tensor input_663_interleave_0 = const()[name = tensor("input_663_interleave_0"), val = tensor(false)]; + tensor input_663 = concat(axis = var_65, interleave = input_663_interleave_0, values = (cache_49, key_25))[name = tensor("input_663")]; + tensor var_2921_begin_0 = const()[name = tensor("op_2921_begin_0"), val = tensor([0, 2, 0])]; + tensor var_2921_end_0 = const()[name = tensor("op_2921_end_0"), val = tensor([1, 70, 1024])]; + tensor var_2921_end_mask_0 = const()[name = tensor("op_2921_end_mask_0"), val = tensor([true, true, true])]; + tensor var_2921 = slice_by_index(begin = var_2921_begin_0, end = var_2921_end_0, end_mask = var_2921_end_mask_0, x = cache_49)[name = tensor("op_2921")]; + tensor var_2927_interleave_0 = const()[name = tensor("op_2927_interleave_0"), val = tensor(false)]; + tensor var_2927 = concat(axis = var_65, interleave = var_2927_interleave_0, values = (var_2921, key_25))[name = tensor("op_2927")]; + tensor var_2930 = linear(bias = linear_2_bias_0, weight = module_layers_12_self_attn_linear_q_weight_quantized, x = key_25)[name = tensor("linear_111")]; + tensor var_2931 = const()[name = tensor("op_2931"), val = tensor([1, -1, 8, 128])]; + tensor q_73 = reshape(shape = var_2931, x = var_2930)[name = tensor("q_73")]; + tensor var_2934 = linear(bias = linear_2_bias_0, weight = module_layers_12_self_attn_linear_k_weight_quantized, x = input_663)[name = tensor("linear_112")]; + tensor var_2935 = const()[name = tensor("op_2935"), val = tensor([1, -1, 8, 128])]; + tensor k_49 = reshape(shape = var_2935, x = var_2934)[name = tensor("k_49")]; + tensor var_2938 = linear(bias = linear_2_bias_0, weight = module_layers_12_self_attn_linear_v_weight_quantized, x = input_663)[name = tensor("linear_113")]; + tensor var_2939 = const()[name = tensor("op_2939"), val = tensor([1, -1, 8, 128])]; + tensor v_25 = reshape(shape = var_2939, x = var_2938)[name = tensor("v_25")]; + tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_2951 = add(x = q_73, y = module_layers_12_self_attn_pos_bias_u)[name = tensor("op_2951")]; + tensor var_2953 = add(x = q_73, y = module_layers_12_self_attn_pos_bias_v)[name = tensor("op_2953")]; + tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_2955_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_2955_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589101888))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589248640))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589248384)))]; + tensor x_319_transpose_x_0 = const()[name = tensor("x_319_transpose_x_0"), val = tensor(false)]; + tensor x_319_transpose_y_0 = const()[name = tensor("x_319_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_25 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2953)[name = tensor("transpose_254")]; + tensor x_319 = matmul(transpose_x = x_319_transpose_x_0, transpose_y = x_319_transpose_y_0, x = q_with_bias_v_25, y = op_2955_quantized)[name = tensor("x_319")]; + tensor const_235 = const()[name = tensor("const_235"), val = tensor(0x0p+0)]; + tensor x_321_pad_0 = const()[name = tensor("x_321_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_321_mode_0 = const()[name = tensor("x_321_mode_0"), val = tensor("constant")]; + tensor x_321 = pad(constant_val = const_235, mode = x_321_mode_0, pad = x_321_pad_0, x = x_319)[name = tensor("x_321")]; + tensor var_2963 = const()[name = tensor("op_2963"), val = tensor([1, 8, -1, 2])]; + tensor x_323 = reshape(shape = var_2963, x = x_321)[name = tensor("x_323")]; + tensor var_2967_begin_0 = const()[name = tensor("op_2967_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_2967_end_0 = const()[name = tensor("op_2967_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_2967_end_mask_0 = const()[name = tensor("op_2967_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_2967 = slice_by_index(begin = var_2967_begin_0, end = var_2967_end_0, end_mask = var_2967_end_mask_0, x = x_323)[name = tensor("op_2967")]; + tensor var_2968 = const()[name = tensor("op_2968"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_49 = reshape(shape = var_2968, x = var_2967)[name = tensor("matrix_bd_49")]; + tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; + tensor transpose_120_perm_0 = const()[name = tensor("transpose_120_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_121_perm_0 = const()[name = tensor("transpose_121_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_121 = transpose(perm = transpose_121_perm_0, x = k_49)[name = tensor("transpose_252")]; + tensor transpose_120 = transpose(perm = transpose_120_perm_0, x = var_2951)[name = tensor("transpose_253")]; + tensor matrix_ac_25 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_120, y = transpose_121)[name = tensor("matrix_ac_25")]; + tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_51 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49)[name = tensor("matrix_bd_51")]; + tensor var_2977 = add(x = matrix_ac_25, y = matrix_bd_51)[name = tensor("op_2977")]; + tensor _inversed_scores_49_y_0 = const()[name = tensor("_inversed_scores_49_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_49 = mul(x = var_2977, y = _inversed_scores_49_y_0)[name = tensor("_inversed_scores_49")]; + tensor scores_51 = select(a = var_41, b = _inversed_scores_49, cond = mask_11)[name = tensor("scores_51")]; + tensor var_2983 = softmax(axis = var_56, x = scores_51)[name = tensor("op_2983")]; + tensor input_665 = select(a = var_40, b = var_2983, cond = mask_11)[name = tensor("input_665")]; + tensor x_325_transpose_x_0 = const()[name = tensor("x_325_transpose_x_0"), val = tensor(false)]; + tensor x_325_transpose_y_0 = const()[name = tensor("x_325_transpose_y_0"), val = tensor(false)]; + tensor value_33 = transpose(perm = value_33_perm_0, x = v_25)[name = tensor("transpose_251")]; + tensor x_325 = matmul(transpose_x = x_325_transpose_x_0, transpose_y = x_325_transpose_y_0, x = input_665, y = value_33)[name = tensor("x_325")]; + tensor var_2987_perm_0 = const()[name = tensor("op_2987_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2988 = const()[name = tensor("op_2988"), val = tensor([1, -1, 1024])]; + tensor var_2987 = transpose(perm = var_2987_perm_0, x = x_325)[name = tensor("transpose_250")]; + tensor input_667 = reshape(shape = var_2988, x = var_2987)[name = tensor("input_667")]; + tensor input_669 = linear(bias = linear_2_bias_0, weight = module_layers_12_self_attn_linear_out_weight_quantized, x = input_667)[name = tensor("linear_115")]; + tensor input_671 = add(x = input_661, y = input_669)[name = tensor("input_671")]; + tensor x_329_axes_0 = const()[name = tensor("x_329_axes_0"), val = tensor([-1])]; + tensor x_329 = layer_norm(axes = x_329_axes_0, beta = module_layers_12_norm_conv_bias, epsilon = var_38, gamma = module_layers_12_norm_conv_weight, x = input_671)[name = tensor("x_329")]; + tensor input_673_perm_0 = const()[name = tensor("input_673_perm_0"), val = tensor([0, 2, 1])]; + tensor input_675_pad_type_0 = const()[name = tensor("input_675_pad_type_0"), val = tensor("valid")]; + tensor input_675_strides_0 = const()[name = tensor("input_675_strides_0"), val = tensor([1])]; + tensor input_675_pad_0 = const()[name = tensor("input_675_pad_0"), val = tensor([0, 0])]; + tensor input_675_dilations_0 = const()[name = tensor("input_675_dilations_0"), val = tensor([1])]; + tensor input_675_groups_0 = const()[name = tensor("input_675_groups_0"), val = tensor(1)]; + tensor input_673 = transpose(perm = input_673_perm_0, x = x_329)[name = tensor("transpose_249")]; + tensor input_675 = conv(dilations = input_675_dilations_0, groups = input_675_groups_0, pad = input_675_pad_0, pad_type = input_675_pad_type_0, strides = input_675_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_quantized, x = input_673)[name = tensor("input_675")]; + tensor x_331_split_num_splits_0 = const()[name = tensor("x_331_split_num_splits_0"), val = tensor(2)]; + tensor x_331_split_axis_0 = const()[name = tensor("x_331_split_axis_0"), val = tensor(1)]; + tensor x_331_split_0, tensor x_331_split_1 = split(axis = x_331_split_axis_0, num_splits = x_331_split_num_splits_0, x = input_675)[name = tensor("x_331_split")]; + tensor x_331_split_1_sigmoid = sigmoid(x = x_331_split_1)[name = tensor("x_331_split_1_sigmoid")]; + tensor x_331 = mul(x = x_331_split_0, y = x_331_split_1_sigmoid)[name = tensor("x_331")]; + tensor input_677 = select(a = var_40, b = x_331, cond = var_565)[name = tensor("input_677")]; + tensor new_x_51_interleave_0 = const()[name = tensor("new_x_51_interleave_0"), val = tensor(false)]; + tensor new_x_51 = concat(axis = var_56, interleave = new_x_51_interleave_0, values = (cache_51, input_677))[name = tensor("new_x_51")]; + tensor var_3026_begin_0 = const()[name = tensor("op_3026_begin_0"), val = tensor([0, 0, 2])]; + tensor var_3026_end_0 = const()[name = tensor("op_3026_end_0"), val = tensor([1, 1024, 10])]; + tensor var_3026_end_mask_0 = const()[name = tensor("op_3026_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3026 = slice_by_index(begin = var_3026_begin_0, end = var_3026_end_0, end_mask = var_3026_end_mask_0, x = new_x_51)[name = tensor("op_3026")]; + tensor x_333_pad_type_0 = const()[name = tensor("x_333_pad_type_0"), val = tensor("valid")]; + tensor x_333_groups_0 = const()[name = tensor("x_333_groups_0"), val = tensor(1024)]; + tensor x_333_strides_0 = const()[name = tensor("x_333_strides_0"), val = tensor([1])]; + tensor x_333_pad_0 = const()[name = tensor("x_333_pad_0"), val = tensor([0, 0])]; + tensor x_333_dilations_0 = const()[name = tensor("x_333_dilations_0"), val = tensor([1])]; + tensor x_333 = conv(dilations = x_333_dilations_0, groups = x_333_groups_0, pad = x_333_pad_0, pad_type = x_333_pad_type_0, strides = x_333_strides_0, weight = module_layers_12_conv_depthwise_conv_weight_quantized, x = new_x_51)[name = tensor("x_333")]; + tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; + tensor x_335_axes_0 = const()[name = tensor("x_335_axes_0"), val = tensor([-1])]; + tensor input_679 = transpose(perm = input_679_perm_0, x = x_333)[name = tensor("transpose_248")]; + tensor x_335 = layer_norm(axes = x_335_axes_0, beta = module_layers_12_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_12_conv_batch_norm_weight, x = input_679)[name = tensor("x_335")]; + tensor input_681_perm_0 = const()[name = tensor("input_681_perm_0"), val = tensor([0, 2, 1])]; + tensor input_681 = transpose(perm = input_681_perm_0, x = x_335)[name = tensor("transpose_247")]; + tensor input_683 = silu(x = input_681)[name = tensor("input_683")]; + tensor x_337_pad_type_0 = const()[name = tensor("x_337_pad_type_0"), val = tensor("valid")]; + tensor x_337_strides_0 = const()[name = tensor("x_337_strides_0"), val = tensor([1])]; + tensor x_337_pad_0 = const()[name = tensor("x_337_pad_0"), val = tensor([0, 0])]; + tensor x_337_dilations_0 = const()[name = tensor("x_337_dilations_0"), val = tensor([1])]; + tensor x_337_groups_0 = const()[name = tensor("x_337_groups_0"), val = tensor(1)]; + tensor x_337 = conv(dilations = x_337_dilations_0, groups = x_337_groups_0, pad = x_337_pad_0, pad_type = x_337_pad_type_0, strides = x_337_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_quantized, x = input_683)[name = tensor("x_337")]; + tensor input_685_perm_0 = const()[name = tensor("input_685_perm_0"), val = tensor([0, 2, 1])]; + tensor input_685 = transpose(perm = input_685_perm_0, x = x_337)[name = tensor("transpose_246")]; + tensor input_687 = add(x = input_671, y = input_685)[name = tensor("input_687")]; + tensor input_689_axes_0 = const()[name = tensor("input_689_axes_0"), val = tensor([-1])]; + tensor input_689 = layer_norm(axes = input_689_axes_0, beta = module_layers_12_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_12_norm_feed_forward2_weight, x = input_687)[name = tensor("input_689")]; + tensor input_691 = linear(bias = linear_1_bias_0, weight = module_layers_12_feed_forward2_linear1_weight_quantized, x = input_689)[name = tensor("linear_116")]; + tensor input_693 = silu(x = input_691)[name = tensor("input_693")]; + tensor input_697 = linear(bias = linear_2_bias_0, weight = module_layers_12_feed_forward2_linear2_weight_quantized, x = input_693)[name = tensor("linear_117")]; + tensor var_3067 = const()[name = tensor("op_3067"), val = tensor(0x1p-1)]; + tensor var_3068 = mul(x = input_697, y = var_3067)[name = tensor("op_3068")]; + tensor input_699 = add(x = input_687, y = var_3068)[name = tensor("input_699")]; + tensor input_701_axes_0 = const()[name = tensor("input_701_axes_0"), val = tensor([-1])]; + tensor input_701 = layer_norm(axes = input_701_axes_0, beta = module_layers_12_norm_out_bias, epsilon = var_38, gamma = module_layers_12_norm_out_weight, x = input_699)[name = tensor("input_701")]; + tensor cache_53_begin_0 = const()[name = tensor("cache_53_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_53_end_0 = const()[name = tensor("cache_53_end_0"), val = tensor([14, 1, 70, 1024])]; + tensor cache_53_end_mask_0 = const()[name = tensor("cache_53_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_53_squeeze_mask_0 = const()[name = tensor("cache_53_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_53 = slice_by_index(begin = cache_53_begin_0, end = cache_53_end_0, end_mask = cache_53_end_mask_0, squeeze_mask = cache_53_squeeze_mask_0, x = value_3)[name = tensor("cache_53")]; + tensor cache_55_begin_0 = const()[name = tensor("cache_55_begin_0"), val = tensor([13, 0, 0, 0])]; + tensor cache_55_end_0 = const()[name = tensor("cache_55_end_0"), val = tensor([14, 1, 1024, 8])]; + tensor cache_55_end_mask_0 = const()[name = tensor("cache_55_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_55_squeeze_mask_0 = const()[name = tensor("cache_55_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_55 = slice_by_index(begin = cache_55_begin_0, end = cache_55_end_0, end_mask = cache_55_end_mask_0, squeeze_mask = cache_55_squeeze_mask_0, x = value_5)[name = tensor("cache_55")]; + tensor input_703_axes_0 = const()[name = tensor("input_703_axes_0"), val = tensor([-1])]; + tensor input_703 = layer_norm(axes = input_703_axes_0, beta = module_layers_13_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_13_norm_feed_forward1_weight, x = input_701)[name = tensor("input_703")]; + tensor input_705 = linear(bias = linear_1_bias_0, weight = module_layers_13_feed_forward1_linear1_weight_quantized, x = input_703)[name = tensor("linear_118")]; + tensor input_707 = silu(x = input_705)[name = tensor("input_707")]; + tensor input_711 = linear(bias = linear_2_bias_0, weight = module_layers_13_feed_forward1_linear2_weight_quantized, x = input_707)[name = tensor("linear_119")]; + tensor var_3102 = const()[name = tensor("op_3102"), val = tensor(0x1p-1)]; + tensor var_3103 = mul(x = input_711, y = var_3102)[name = tensor("op_3103")]; + tensor input_713 = add(x = input_701, y = var_3103)[name = tensor("input_713")]; + tensor key_27_axes_0 = const()[name = tensor("key_27_axes_0"), val = tensor([-1])]; + tensor key_27 = layer_norm(axes = key_27_axes_0, beta = module_layers_13_norm_self_att_bias, epsilon = var_38, gamma = module_layers_13_norm_self_att_weight, x = input_713)[name = tensor("key_27")]; + tensor input_715_interleave_0 = const()[name = tensor("input_715_interleave_0"), val = tensor(false)]; + tensor input_715 = concat(axis = var_65, interleave = input_715_interleave_0, values = (cache_53, key_27))[name = tensor("input_715")]; + tensor var_3125_begin_0 = const()[name = tensor("op_3125_begin_0"), val = tensor([0, 2, 0])]; + tensor var_3125_end_0 = const()[name = tensor("op_3125_end_0"), val = tensor([1, 70, 1024])]; + tensor var_3125_end_mask_0 = const()[name = tensor("op_3125_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3125 = slice_by_index(begin = var_3125_begin_0, end = var_3125_end_0, end_mask = var_3125_end_mask_0, x = cache_53)[name = tensor("op_3125")]; + tensor var_3131_interleave_0 = const()[name = tensor("op_3131_interleave_0"), val = tensor(false)]; + tensor var_3131 = concat(axis = var_65, interleave = var_3131_interleave_0, values = (var_3125, key_27))[name = tensor("op_3131")]; + tensor var_3134 = linear(bias = linear_2_bias_0, weight = module_layers_13_self_attn_linear_q_weight_quantized, x = key_27)[name = tensor("linear_120")]; + tensor var_3135 = const()[name = tensor("op_3135"), val = tensor([1, -1, 8, 128])]; + tensor q_79 = reshape(shape = var_3135, x = var_3134)[name = tensor("q_79")]; + tensor var_3138 = linear(bias = linear_2_bias_0, weight = module_layers_13_self_attn_linear_k_weight_quantized, x = input_715)[name = tensor("linear_121")]; + tensor var_3139 = const()[name = tensor("op_3139"), val = tensor([1, -1, 8, 128])]; + tensor k_53 = reshape(shape = var_3139, x = var_3138)[name = tensor("k_53")]; + tensor var_3142 = linear(bias = linear_2_bias_0, weight = module_layers_13_self_attn_linear_v_weight_quantized, x = input_715)[name = tensor("linear_122")]; + tensor var_3143 = const()[name = tensor("op_3143"), val = tensor([1, -1, 8, 128])]; + tensor v_27 = reshape(shape = var_3143, x = var_3142)[name = tensor("v_27")]; + tensor value_35_perm_0 = const()[name = tensor("value_35_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_3155 = add(x = q_79, y = module_layers_13_self_attn_pos_bias_u)[name = tensor("op_3155")]; + tensor var_3157 = add(x = q_79, y = module_layers_13_self_attn_pos_bias_v)[name = tensor("op_3157")]; + tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_3159_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3159_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589249280))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589396032))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589395776)))]; + tensor x_345_transpose_x_0 = const()[name = tensor("x_345_transpose_x_0"), val = tensor(false)]; + tensor x_345_transpose_y_0 = const()[name = tensor("x_345_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_27 = transpose(perm = q_with_bias_v_27_perm_0, x = var_3157)[name = tensor("transpose_245")]; + tensor x_345 = matmul(transpose_x = x_345_transpose_x_0, transpose_y = x_345_transpose_y_0, x = q_with_bias_v_27, y = op_3159_quantized)[name = tensor("x_345")]; + tensor const_248 = const()[name = tensor("const_248"), val = tensor(0x0p+0)]; + tensor x_347_pad_0 = const()[name = tensor("x_347_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_347_mode_0 = const()[name = tensor("x_347_mode_0"), val = tensor("constant")]; + tensor x_347 = pad(constant_val = const_248, mode = x_347_mode_0, pad = x_347_pad_0, x = x_345)[name = tensor("x_347")]; + tensor var_3167 = const()[name = tensor("op_3167"), val = tensor([1, 8, -1, 2])]; + tensor x_349 = reshape(shape = var_3167, x = x_347)[name = tensor("x_349")]; + tensor var_3171_begin_0 = const()[name = tensor("op_3171_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3171_end_0 = const()[name = tensor("op_3171_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_3171_end_mask_0 = const()[name = tensor("op_3171_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3171 = slice_by_index(begin = var_3171_begin_0, end = var_3171_end_0, end_mask = var_3171_end_mask_0, x = x_349)[name = tensor("op_3171")]; + tensor var_3172 = const()[name = tensor("op_3172"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_53 = reshape(shape = var_3172, x = var_3171)[name = tensor("matrix_bd_53")]; + tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; + tensor transpose_122_perm_0 = const()[name = tensor("transpose_122_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_123_perm_0 = const()[name = tensor("transpose_123_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_123 = transpose(perm = transpose_123_perm_0, x = k_53)[name = tensor("transpose_243")]; + tensor transpose_122 = transpose(perm = transpose_122_perm_0, x = var_3155)[name = tensor("transpose_244")]; + tensor matrix_ac_27 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_122, y = transpose_123)[name = tensor("matrix_ac_27")]; + tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_55 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53)[name = tensor("matrix_bd_55")]; + tensor var_3181 = add(x = matrix_ac_27, y = matrix_bd_55)[name = tensor("op_3181")]; + tensor _inversed_scores_53_y_0 = const()[name = tensor("_inversed_scores_53_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_53 = mul(x = var_3181, y = _inversed_scores_53_y_0)[name = tensor("_inversed_scores_53")]; + tensor scores_55 = select(a = var_41, b = _inversed_scores_53, cond = mask_11)[name = tensor("scores_55")]; + tensor var_3187 = softmax(axis = var_56, x = scores_55)[name = tensor("op_3187")]; + tensor input_717 = select(a = var_40, b = var_3187, cond = mask_11)[name = tensor("input_717")]; + tensor x_351_transpose_x_0 = const()[name = tensor("x_351_transpose_x_0"), val = tensor(false)]; + tensor x_351_transpose_y_0 = const()[name = tensor("x_351_transpose_y_0"), val = tensor(false)]; + tensor value_35 = transpose(perm = value_35_perm_0, x = v_27)[name = tensor("transpose_242")]; + tensor x_351 = matmul(transpose_x = x_351_transpose_x_0, transpose_y = x_351_transpose_y_0, x = input_717, y = value_35)[name = tensor("x_351")]; + tensor var_3191_perm_0 = const()[name = tensor("op_3191_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3192 = const()[name = tensor("op_3192"), val = tensor([1, -1, 1024])]; + tensor var_3191 = transpose(perm = var_3191_perm_0, x = x_351)[name = tensor("transpose_241")]; + tensor input_719 = reshape(shape = var_3192, x = var_3191)[name = tensor("input_719")]; + tensor input_721 = linear(bias = linear_2_bias_0, weight = module_layers_13_self_attn_linear_out_weight_quantized, x = input_719)[name = tensor("linear_124")]; + tensor input_723 = add(x = input_713, y = input_721)[name = tensor("input_723")]; + tensor x_355_axes_0 = const()[name = tensor("x_355_axes_0"), val = tensor([-1])]; + tensor x_355 = layer_norm(axes = x_355_axes_0, beta = module_layers_13_norm_conv_bias, epsilon = var_38, gamma = module_layers_13_norm_conv_weight, x = input_723)[name = tensor("x_355")]; + tensor input_725_perm_0 = const()[name = tensor("input_725_perm_0"), val = tensor([0, 2, 1])]; + tensor input_727_pad_type_0 = const()[name = tensor("input_727_pad_type_0"), val = tensor("valid")]; + tensor input_727_strides_0 = const()[name = tensor("input_727_strides_0"), val = tensor([1])]; + tensor input_727_pad_0 = const()[name = tensor("input_727_pad_0"), val = tensor([0, 0])]; + tensor input_727_dilations_0 = const()[name = tensor("input_727_dilations_0"), val = tensor([1])]; + tensor input_727_groups_0 = const()[name = tensor("input_727_groups_0"), val = tensor(1)]; + tensor input_725 = transpose(perm = input_725_perm_0, x = x_355)[name = tensor("transpose_240")]; + tensor input_727 = conv(dilations = input_727_dilations_0, groups = input_727_groups_0, pad = input_727_pad_0, pad_type = input_727_pad_type_0, strides = input_727_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_quantized, x = input_725)[name = tensor("input_727")]; + tensor x_357_split_num_splits_0 = const()[name = tensor("x_357_split_num_splits_0"), val = tensor(2)]; + tensor x_357_split_axis_0 = const()[name = tensor("x_357_split_axis_0"), val = tensor(1)]; + tensor x_357_split_0, tensor x_357_split_1 = split(axis = x_357_split_axis_0, num_splits = x_357_split_num_splits_0, x = input_727)[name = tensor("x_357_split")]; + tensor x_357_split_1_sigmoid = sigmoid(x = x_357_split_1)[name = tensor("x_357_split_1_sigmoid")]; + tensor x_357 = mul(x = x_357_split_0, y = x_357_split_1_sigmoid)[name = tensor("x_357")]; + tensor input_729 = select(a = var_40, b = x_357, cond = var_565)[name = tensor("input_729")]; + tensor new_x_55_interleave_0 = const()[name = tensor("new_x_55_interleave_0"), val = tensor(false)]; + tensor new_x_55 = concat(axis = var_56, interleave = new_x_55_interleave_0, values = (cache_55, input_729))[name = tensor("new_x_55")]; + tensor var_3230_begin_0 = const()[name = tensor("op_3230_begin_0"), val = tensor([0, 0, 2])]; + tensor var_3230_end_0 = const()[name = tensor("op_3230_end_0"), val = tensor([1, 1024, 10])]; + tensor var_3230_end_mask_0 = const()[name = tensor("op_3230_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3230 = slice_by_index(begin = var_3230_begin_0, end = var_3230_end_0, end_mask = var_3230_end_mask_0, x = new_x_55)[name = tensor("op_3230")]; + tensor x_359_pad_type_0 = const()[name = tensor("x_359_pad_type_0"), val = tensor("valid")]; + tensor x_359_groups_0 = const()[name = tensor("x_359_groups_0"), val = tensor(1024)]; + tensor x_359_strides_0 = const()[name = tensor("x_359_strides_0"), val = tensor([1])]; + tensor x_359_pad_0 = const()[name = tensor("x_359_pad_0"), val = tensor([0, 0])]; + tensor x_359_dilations_0 = const()[name = tensor("x_359_dilations_0"), val = tensor([1])]; + tensor x_359 = conv(dilations = x_359_dilations_0, groups = x_359_groups_0, pad = x_359_pad_0, pad_type = x_359_pad_type_0, strides = x_359_strides_0, weight = module_layers_13_conv_depthwise_conv_weight_quantized, x = new_x_55)[name = tensor("x_359")]; + tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; + tensor x_361_axes_0 = const()[name = tensor("x_361_axes_0"), val = tensor([-1])]; + tensor input_731 = transpose(perm = input_731_perm_0, x = x_359)[name = tensor("transpose_239")]; + tensor x_361 = layer_norm(axes = x_361_axes_0, beta = module_layers_13_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_13_conv_batch_norm_weight, x = input_731)[name = tensor("x_361")]; + tensor input_733_perm_0 = const()[name = tensor("input_733_perm_0"), val = tensor([0, 2, 1])]; + tensor input_733 = transpose(perm = input_733_perm_0, x = x_361)[name = tensor("transpose_238")]; + tensor input_735 = silu(x = input_733)[name = tensor("input_735")]; + tensor x_363_pad_type_0 = const()[name = tensor("x_363_pad_type_0"), val = tensor("valid")]; + tensor x_363_strides_0 = const()[name = tensor("x_363_strides_0"), val = tensor([1])]; + tensor x_363_pad_0 = const()[name = tensor("x_363_pad_0"), val = tensor([0, 0])]; + tensor x_363_dilations_0 = const()[name = tensor("x_363_dilations_0"), val = tensor([1])]; + tensor x_363_groups_0 = const()[name = tensor("x_363_groups_0"), val = tensor(1)]; + tensor x_363 = conv(dilations = x_363_dilations_0, groups = x_363_groups_0, pad = x_363_pad_0, pad_type = x_363_pad_type_0, strides = x_363_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_quantized, x = input_735)[name = tensor("x_363")]; + tensor input_737_perm_0 = const()[name = tensor("input_737_perm_0"), val = tensor([0, 2, 1])]; + tensor input_737 = transpose(perm = input_737_perm_0, x = x_363)[name = tensor("transpose_237")]; + tensor input_739 = add(x = input_723, y = input_737)[name = tensor("input_739")]; + tensor input_741_axes_0 = const()[name = tensor("input_741_axes_0"), val = tensor([-1])]; + tensor input_741 = layer_norm(axes = input_741_axes_0, beta = module_layers_13_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_13_norm_feed_forward2_weight, x = input_739)[name = tensor("input_741")]; + tensor input_743 = linear(bias = linear_1_bias_0, weight = module_layers_13_feed_forward2_linear1_weight_quantized, x = input_741)[name = tensor("linear_125")]; + tensor input_745 = silu(x = input_743)[name = tensor("input_745")]; + tensor input_749 = linear(bias = linear_2_bias_0, weight = module_layers_13_feed_forward2_linear2_weight_quantized, x = input_745)[name = tensor("linear_126")]; + tensor var_3271 = const()[name = tensor("op_3271"), val = tensor(0x1p-1)]; + tensor var_3272 = mul(x = input_749, y = var_3271)[name = tensor("op_3272")]; + tensor input_751 = add(x = input_739, y = var_3272)[name = tensor("input_751")]; + tensor input_753_axes_0 = const()[name = tensor("input_753_axes_0"), val = tensor([-1])]; + tensor input_753 = layer_norm(axes = input_753_axes_0, beta = module_layers_13_norm_out_bias, epsilon = var_38, gamma = module_layers_13_norm_out_weight, x = input_751)[name = tensor("input_753")]; + tensor cache_57_begin_0 = const()[name = tensor("cache_57_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_57_end_0 = const()[name = tensor("cache_57_end_0"), val = tensor([15, 1, 70, 1024])]; + tensor cache_57_end_mask_0 = const()[name = tensor("cache_57_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_57_squeeze_mask_0 = const()[name = tensor("cache_57_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_57 = slice_by_index(begin = cache_57_begin_0, end = cache_57_end_0, end_mask = cache_57_end_mask_0, squeeze_mask = cache_57_squeeze_mask_0, x = value_3)[name = tensor("cache_57")]; + tensor cache_59_begin_0 = const()[name = tensor("cache_59_begin_0"), val = tensor([14, 0, 0, 0])]; + tensor cache_59_end_0 = const()[name = tensor("cache_59_end_0"), val = tensor([15, 1, 1024, 8])]; + tensor cache_59_end_mask_0 = const()[name = tensor("cache_59_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_59_squeeze_mask_0 = const()[name = tensor("cache_59_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_59 = slice_by_index(begin = cache_59_begin_0, end = cache_59_end_0, end_mask = cache_59_end_mask_0, squeeze_mask = cache_59_squeeze_mask_0, x = value_5)[name = tensor("cache_59")]; + tensor input_755_axes_0 = const()[name = tensor("input_755_axes_0"), val = tensor([-1])]; + tensor input_755 = layer_norm(axes = input_755_axes_0, beta = module_layers_14_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_14_norm_feed_forward1_weight, x = input_753)[name = tensor("input_755")]; + tensor input_757 = linear(bias = linear_1_bias_0, weight = module_layers_14_feed_forward1_linear1_weight_quantized, x = input_755)[name = tensor("linear_127")]; + tensor input_759 = silu(x = input_757)[name = tensor("input_759")]; + tensor input_763 = linear(bias = linear_2_bias_0, weight = module_layers_14_feed_forward1_linear2_weight_quantized, x = input_759)[name = tensor("linear_128")]; + tensor var_3306 = const()[name = tensor("op_3306"), val = tensor(0x1p-1)]; + tensor var_3307 = mul(x = input_763, y = var_3306)[name = tensor("op_3307")]; + tensor input_765 = add(x = input_753, y = var_3307)[name = tensor("input_765")]; + tensor key_29_axes_0 = const()[name = tensor("key_29_axes_0"), val = tensor([-1])]; + tensor key_29 = layer_norm(axes = key_29_axes_0, beta = module_layers_14_norm_self_att_bias, epsilon = var_38, gamma = module_layers_14_norm_self_att_weight, x = input_765)[name = tensor("key_29")]; + tensor input_767_interleave_0 = const()[name = tensor("input_767_interleave_0"), val = tensor(false)]; + tensor input_767 = concat(axis = var_65, interleave = input_767_interleave_0, values = (cache_57, key_29))[name = tensor("input_767")]; + tensor var_3329_begin_0 = const()[name = tensor("op_3329_begin_0"), val = tensor([0, 2, 0])]; + tensor var_3329_end_0 = const()[name = tensor("op_3329_end_0"), val = tensor([1, 70, 1024])]; + tensor var_3329_end_mask_0 = const()[name = tensor("op_3329_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3329 = slice_by_index(begin = var_3329_begin_0, end = var_3329_end_0, end_mask = var_3329_end_mask_0, x = cache_57)[name = tensor("op_3329")]; + tensor var_3335_interleave_0 = const()[name = tensor("op_3335_interleave_0"), val = tensor(false)]; + tensor var_3335 = concat(axis = var_65, interleave = var_3335_interleave_0, values = (var_3329, key_29))[name = tensor("op_3335")]; + tensor var_3338 = linear(bias = linear_2_bias_0, weight = module_layers_14_self_attn_linear_q_weight_quantized, x = key_29)[name = tensor("linear_129")]; + tensor var_3339 = const()[name = tensor("op_3339"), val = tensor([1, -1, 8, 128])]; + tensor q_85 = reshape(shape = var_3339, x = var_3338)[name = tensor("q_85")]; + tensor var_3342 = linear(bias = linear_2_bias_0, weight = module_layers_14_self_attn_linear_k_weight_quantized, x = input_767)[name = tensor("linear_130")]; + tensor var_3343 = const()[name = tensor("op_3343"), val = tensor([1, -1, 8, 128])]; + tensor k_57 = reshape(shape = var_3343, x = var_3342)[name = tensor("k_57")]; + tensor var_3346 = linear(bias = linear_2_bias_0, weight = module_layers_14_self_attn_linear_v_weight_quantized, x = input_767)[name = tensor("linear_131")]; + tensor var_3347 = const()[name = tensor("op_3347"), val = tensor([1, -1, 8, 128])]; + tensor v_29 = reshape(shape = var_3347, x = var_3346)[name = tensor("v_29")]; + tensor value_37_perm_0 = const()[name = tensor("value_37_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_3359 = add(x = q_85, y = module_layers_14_self_attn_pos_bias_u)[name = tensor("op_3359")]; + tensor var_3361 = add(x = q_85, y = module_layers_14_self_attn_pos_bias_v)[name = tensor("op_3361")]; + tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_3363_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3363_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589396672))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589543424))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589543168)))]; + tensor x_371_transpose_x_0 = const()[name = tensor("x_371_transpose_x_0"), val = tensor(false)]; + tensor x_371_transpose_y_0 = const()[name = tensor("x_371_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_29 = transpose(perm = q_with_bias_v_29_perm_0, x = var_3361)[name = tensor("transpose_236")]; + tensor x_371 = matmul(transpose_x = x_371_transpose_x_0, transpose_y = x_371_transpose_y_0, x = q_with_bias_v_29, y = op_3363_quantized)[name = tensor("x_371")]; + tensor const_261 = const()[name = tensor("const_261"), val = tensor(0x0p+0)]; + tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_373_mode_0 = const()[name = tensor("x_373_mode_0"), val = tensor("constant")]; + tensor x_373 = pad(constant_val = const_261, mode = x_373_mode_0, pad = x_373_pad_0, x = x_371)[name = tensor("x_373")]; + tensor var_3371 = const()[name = tensor("op_3371"), val = tensor([1, 8, -1, 2])]; + tensor x_375 = reshape(shape = var_3371, x = x_373)[name = tensor("x_375")]; + tensor var_3375_begin_0 = const()[name = tensor("op_3375_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3375_end_0 = const()[name = tensor("op_3375_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_3375_end_mask_0 = const()[name = tensor("op_3375_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3375 = slice_by_index(begin = var_3375_begin_0, end = var_3375_end_0, end_mask = var_3375_end_mask_0, x = x_375)[name = tensor("op_3375")]; + tensor var_3376 = const()[name = tensor("op_3376"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_57 = reshape(shape = var_3376, x = var_3375)[name = tensor("matrix_bd_57")]; + tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; + tensor transpose_124_perm_0 = const()[name = tensor("transpose_124_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_125_perm_0 = const()[name = tensor("transpose_125_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_125 = transpose(perm = transpose_125_perm_0, x = k_57)[name = tensor("transpose_234")]; + tensor transpose_124 = transpose(perm = transpose_124_perm_0, x = var_3359)[name = tensor("transpose_235")]; + tensor matrix_ac_29 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_124, y = transpose_125)[name = tensor("matrix_ac_29")]; + tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_59 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57)[name = tensor("matrix_bd_59")]; + tensor var_3385 = add(x = matrix_ac_29, y = matrix_bd_59)[name = tensor("op_3385")]; + tensor _inversed_scores_57_y_0 = const()[name = tensor("_inversed_scores_57_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_57 = mul(x = var_3385, y = _inversed_scores_57_y_0)[name = tensor("_inversed_scores_57")]; + tensor scores_59 = select(a = var_41, b = _inversed_scores_57, cond = mask_11)[name = tensor("scores_59")]; + tensor var_3391 = softmax(axis = var_56, x = scores_59)[name = tensor("op_3391")]; + tensor input_769 = select(a = var_40, b = var_3391, cond = mask_11)[name = tensor("input_769")]; + tensor x_377_transpose_x_0 = const()[name = tensor("x_377_transpose_x_0"), val = tensor(false)]; + tensor x_377_transpose_y_0 = const()[name = tensor("x_377_transpose_y_0"), val = tensor(false)]; + tensor value_37 = transpose(perm = value_37_perm_0, x = v_29)[name = tensor("transpose_233")]; + tensor x_377 = matmul(transpose_x = x_377_transpose_x_0, transpose_y = x_377_transpose_y_0, x = input_769, y = value_37)[name = tensor("x_377")]; + tensor var_3395_perm_0 = const()[name = tensor("op_3395_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3396 = const()[name = tensor("op_3396"), val = tensor([1, -1, 1024])]; + tensor var_3395 = transpose(perm = var_3395_perm_0, x = x_377)[name = tensor("transpose_232")]; + tensor input_771 = reshape(shape = var_3396, x = var_3395)[name = tensor("input_771")]; + tensor input_773 = linear(bias = linear_2_bias_0, weight = module_layers_14_self_attn_linear_out_weight_quantized, x = input_771)[name = tensor("linear_133")]; + tensor input_775 = add(x = input_765, y = input_773)[name = tensor("input_775")]; + tensor x_381_axes_0 = const()[name = tensor("x_381_axes_0"), val = tensor([-1])]; + tensor x_381 = layer_norm(axes = x_381_axes_0, beta = module_layers_14_norm_conv_bias, epsilon = var_38, gamma = module_layers_14_norm_conv_weight, x = input_775)[name = tensor("x_381")]; + tensor input_777_perm_0 = const()[name = tensor("input_777_perm_0"), val = tensor([0, 2, 1])]; + tensor input_779_pad_type_0 = const()[name = tensor("input_779_pad_type_0"), val = tensor("valid")]; + tensor input_779_strides_0 = const()[name = tensor("input_779_strides_0"), val = tensor([1])]; + tensor input_779_pad_0 = const()[name = tensor("input_779_pad_0"), val = tensor([0, 0])]; + tensor input_779_dilations_0 = const()[name = tensor("input_779_dilations_0"), val = tensor([1])]; + tensor input_779_groups_0 = const()[name = tensor("input_779_groups_0"), val = tensor(1)]; + tensor input_777 = transpose(perm = input_777_perm_0, x = x_381)[name = tensor("transpose_231")]; + tensor input_779 = conv(dilations = input_779_dilations_0, groups = input_779_groups_0, pad = input_779_pad_0, pad_type = input_779_pad_type_0, strides = input_779_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_quantized, x = input_777)[name = tensor("input_779")]; + tensor x_383_split_num_splits_0 = const()[name = tensor("x_383_split_num_splits_0"), val = tensor(2)]; + tensor x_383_split_axis_0 = const()[name = tensor("x_383_split_axis_0"), val = tensor(1)]; + tensor x_383_split_0, tensor x_383_split_1 = split(axis = x_383_split_axis_0, num_splits = x_383_split_num_splits_0, x = input_779)[name = tensor("x_383_split")]; + tensor x_383_split_1_sigmoid = sigmoid(x = x_383_split_1)[name = tensor("x_383_split_1_sigmoid")]; + tensor x_383 = mul(x = x_383_split_0, y = x_383_split_1_sigmoid)[name = tensor("x_383")]; + tensor input_781 = select(a = var_40, b = x_383, cond = var_565)[name = tensor("input_781")]; + tensor new_x_59_interleave_0 = const()[name = tensor("new_x_59_interleave_0"), val = tensor(false)]; + tensor new_x_59 = concat(axis = var_56, interleave = new_x_59_interleave_0, values = (cache_59, input_781))[name = tensor("new_x_59")]; + tensor var_3434_begin_0 = const()[name = tensor("op_3434_begin_0"), val = tensor([0, 0, 2])]; + tensor var_3434_end_0 = const()[name = tensor("op_3434_end_0"), val = tensor([1, 1024, 10])]; + tensor var_3434_end_mask_0 = const()[name = tensor("op_3434_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3434 = slice_by_index(begin = var_3434_begin_0, end = var_3434_end_0, end_mask = var_3434_end_mask_0, x = new_x_59)[name = tensor("op_3434")]; + tensor x_385_pad_type_0 = const()[name = tensor("x_385_pad_type_0"), val = tensor("valid")]; + tensor x_385_groups_0 = const()[name = tensor("x_385_groups_0"), val = tensor(1024)]; + tensor x_385_strides_0 = const()[name = tensor("x_385_strides_0"), val = tensor([1])]; + tensor x_385_pad_0 = const()[name = tensor("x_385_pad_0"), val = tensor([0, 0])]; + tensor x_385_dilations_0 = const()[name = tensor("x_385_dilations_0"), val = tensor([1])]; + tensor x_385 = conv(dilations = x_385_dilations_0, groups = x_385_groups_0, pad = x_385_pad_0, pad_type = x_385_pad_type_0, strides = x_385_strides_0, weight = module_layers_14_conv_depthwise_conv_weight_quantized, x = new_x_59)[name = tensor("x_385")]; + tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; + tensor x_387_axes_0 = const()[name = tensor("x_387_axes_0"), val = tensor([-1])]; + tensor input_783 = transpose(perm = input_783_perm_0, x = x_385)[name = tensor("transpose_230")]; + tensor x_387 = layer_norm(axes = x_387_axes_0, beta = module_layers_14_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_14_conv_batch_norm_weight, x = input_783)[name = tensor("x_387")]; + tensor input_785_perm_0 = const()[name = tensor("input_785_perm_0"), val = tensor([0, 2, 1])]; + tensor input_785 = transpose(perm = input_785_perm_0, x = x_387)[name = tensor("transpose_229")]; + tensor input_787 = silu(x = input_785)[name = tensor("input_787")]; + tensor x_389_pad_type_0 = const()[name = tensor("x_389_pad_type_0"), val = tensor("valid")]; + tensor x_389_strides_0 = const()[name = tensor("x_389_strides_0"), val = tensor([1])]; + tensor x_389_pad_0 = const()[name = tensor("x_389_pad_0"), val = tensor([0, 0])]; + tensor x_389_dilations_0 = const()[name = tensor("x_389_dilations_0"), val = tensor([1])]; + tensor x_389_groups_0 = const()[name = tensor("x_389_groups_0"), val = tensor(1)]; + tensor x_389 = conv(dilations = x_389_dilations_0, groups = x_389_groups_0, pad = x_389_pad_0, pad_type = x_389_pad_type_0, strides = x_389_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_quantized, x = input_787)[name = tensor("x_389")]; + tensor input_789_perm_0 = const()[name = tensor("input_789_perm_0"), val = tensor([0, 2, 1])]; + tensor input_789 = transpose(perm = input_789_perm_0, x = x_389)[name = tensor("transpose_228")]; + tensor input_791 = add(x = input_775, y = input_789)[name = tensor("input_791")]; + tensor input_793_axes_0 = const()[name = tensor("input_793_axes_0"), val = tensor([-1])]; + tensor input_793 = layer_norm(axes = input_793_axes_0, beta = module_layers_14_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_14_norm_feed_forward2_weight, x = input_791)[name = tensor("input_793")]; + tensor input_795 = linear(bias = linear_1_bias_0, weight = module_layers_14_feed_forward2_linear1_weight_quantized, x = input_793)[name = tensor("linear_134")]; + tensor input_797 = silu(x = input_795)[name = tensor("input_797")]; + tensor input_801 = linear(bias = linear_2_bias_0, weight = module_layers_14_feed_forward2_linear2_weight_quantized, x = input_797)[name = tensor("linear_135")]; + tensor var_3475 = const()[name = tensor("op_3475"), val = tensor(0x1p-1)]; + tensor var_3476 = mul(x = input_801, y = var_3475)[name = tensor("op_3476")]; + tensor input_803 = add(x = input_791, y = var_3476)[name = tensor("input_803")]; + tensor input_805_axes_0 = const()[name = tensor("input_805_axes_0"), val = tensor([-1])]; + tensor input_805 = layer_norm(axes = input_805_axes_0, beta = module_layers_14_norm_out_bias, epsilon = var_38, gamma = module_layers_14_norm_out_weight, x = input_803)[name = tensor("input_805")]; + tensor cache_61_begin_0 = const()[name = tensor("cache_61_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_61_end_0 = const()[name = tensor("cache_61_end_0"), val = tensor([16, 1, 70, 1024])]; + tensor cache_61_end_mask_0 = const()[name = tensor("cache_61_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_61_squeeze_mask_0 = const()[name = tensor("cache_61_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_61 = slice_by_index(begin = cache_61_begin_0, end = cache_61_end_0, end_mask = cache_61_end_mask_0, squeeze_mask = cache_61_squeeze_mask_0, x = value_3)[name = tensor("cache_61")]; + tensor cache_63_begin_0 = const()[name = tensor("cache_63_begin_0"), val = tensor([15, 0, 0, 0])]; + tensor cache_63_end_0 = const()[name = tensor("cache_63_end_0"), val = tensor([16, 1, 1024, 8])]; + tensor cache_63_end_mask_0 = const()[name = tensor("cache_63_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_63_squeeze_mask_0 = const()[name = tensor("cache_63_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_63 = slice_by_index(begin = cache_63_begin_0, end = cache_63_end_0, end_mask = cache_63_end_mask_0, squeeze_mask = cache_63_squeeze_mask_0, x = value_5)[name = tensor("cache_63")]; + tensor input_807_axes_0 = const()[name = tensor("input_807_axes_0"), val = tensor([-1])]; + tensor input_807 = layer_norm(axes = input_807_axes_0, beta = module_layers_15_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_15_norm_feed_forward1_weight, x = input_805)[name = tensor("input_807")]; + tensor input_809 = linear(bias = linear_1_bias_0, weight = module_layers_15_feed_forward1_linear1_weight_quantized, x = input_807)[name = tensor("linear_136")]; + tensor input_811 = silu(x = input_809)[name = tensor("input_811")]; + tensor input_815 = linear(bias = linear_2_bias_0, weight = module_layers_15_feed_forward1_linear2_weight_quantized, x = input_811)[name = tensor("linear_137")]; + tensor var_3510 = const()[name = tensor("op_3510"), val = tensor(0x1p-1)]; + tensor var_3511 = mul(x = input_815, y = var_3510)[name = tensor("op_3511")]; + tensor input_817 = add(x = input_805, y = var_3511)[name = tensor("input_817")]; + tensor key_31_axes_0 = const()[name = tensor("key_31_axes_0"), val = tensor([-1])]; + tensor key_31 = layer_norm(axes = key_31_axes_0, beta = module_layers_15_norm_self_att_bias, epsilon = var_38, gamma = module_layers_15_norm_self_att_weight, x = input_817)[name = tensor("key_31")]; + tensor input_819_interleave_0 = const()[name = tensor("input_819_interleave_0"), val = tensor(false)]; + tensor input_819 = concat(axis = var_65, interleave = input_819_interleave_0, values = (cache_61, key_31))[name = tensor("input_819")]; + tensor var_3533_begin_0 = const()[name = tensor("op_3533_begin_0"), val = tensor([0, 2, 0])]; + tensor var_3533_end_0 = const()[name = tensor("op_3533_end_0"), val = tensor([1, 70, 1024])]; + tensor var_3533_end_mask_0 = const()[name = tensor("op_3533_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3533 = slice_by_index(begin = var_3533_begin_0, end = var_3533_end_0, end_mask = var_3533_end_mask_0, x = cache_61)[name = tensor("op_3533")]; + tensor var_3539_interleave_0 = const()[name = tensor("op_3539_interleave_0"), val = tensor(false)]; + tensor var_3539 = concat(axis = var_65, interleave = var_3539_interleave_0, values = (var_3533, key_31))[name = tensor("op_3539")]; + tensor var_3542 = linear(bias = linear_2_bias_0, weight = module_layers_15_self_attn_linear_q_weight_quantized, x = key_31)[name = tensor("linear_138")]; + tensor var_3543 = const()[name = tensor("op_3543"), val = tensor([1, -1, 8, 128])]; + tensor q_91 = reshape(shape = var_3543, x = var_3542)[name = tensor("q_91")]; + tensor var_3546 = linear(bias = linear_2_bias_0, weight = module_layers_15_self_attn_linear_k_weight_quantized, x = input_819)[name = tensor("linear_139")]; + tensor var_3547 = const()[name = tensor("op_3547"), val = tensor([1, -1, 8, 128])]; + tensor k_61 = reshape(shape = var_3547, x = var_3546)[name = tensor("k_61")]; + tensor var_3550 = linear(bias = linear_2_bias_0, weight = module_layers_15_self_attn_linear_v_weight_quantized, x = input_819)[name = tensor("linear_140")]; + tensor var_3551 = const()[name = tensor("op_3551"), val = tensor([1, -1, 8, 128])]; + tensor v_31 = reshape(shape = var_3551, x = var_3550)[name = tensor("v_31")]; + tensor value_39_perm_0 = const()[name = tensor("value_39_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_3563 = add(x = q_91, y = module_layers_15_self_attn_pos_bias_u)[name = tensor("op_3563")]; + tensor var_3565 = add(x = q_91, y = module_layers_15_self_attn_pos_bias_v)[name = tensor("op_3565")]; + tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_3567_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3567_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589544064))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589690816))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589690560)))]; + tensor x_397_transpose_x_0 = const()[name = tensor("x_397_transpose_x_0"), val = tensor(false)]; + tensor x_397_transpose_y_0 = const()[name = tensor("x_397_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_31 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3565)[name = tensor("transpose_227")]; + tensor x_397 = matmul(transpose_x = x_397_transpose_x_0, transpose_y = x_397_transpose_y_0, x = q_with_bias_v_31, y = op_3567_quantized)[name = tensor("x_397")]; + tensor const_274 = const()[name = tensor("const_274"), val = tensor(0x0p+0)]; + tensor x_399_pad_0 = const()[name = tensor("x_399_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_399_mode_0 = const()[name = tensor("x_399_mode_0"), val = tensor("constant")]; + tensor x_399 = pad(constant_val = const_274, mode = x_399_mode_0, pad = x_399_pad_0, x = x_397)[name = tensor("x_399")]; + tensor var_3575 = const()[name = tensor("op_3575"), val = tensor([1, 8, -1, 2])]; + tensor x_401 = reshape(shape = var_3575, x = x_399)[name = tensor("x_401")]; + tensor var_3579_begin_0 = const()[name = tensor("op_3579_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3579_end_0 = const()[name = tensor("op_3579_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_3579_end_mask_0 = const()[name = tensor("op_3579_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3579 = slice_by_index(begin = var_3579_begin_0, end = var_3579_end_0, end_mask = var_3579_end_mask_0, x = x_401)[name = tensor("op_3579")]; + tensor var_3580 = const()[name = tensor("op_3580"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_61 = reshape(shape = var_3580, x = var_3579)[name = tensor("matrix_bd_61")]; + tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; + tensor transpose_126_perm_0 = const()[name = tensor("transpose_126_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_127_perm_0 = const()[name = tensor("transpose_127_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_127 = transpose(perm = transpose_127_perm_0, x = k_61)[name = tensor("transpose_225")]; + tensor transpose_126 = transpose(perm = transpose_126_perm_0, x = var_3563)[name = tensor("transpose_226")]; + tensor matrix_ac_31 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_126, y = transpose_127)[name = tensor("matrix_ac_31")]; + tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_63 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61)[name = tensor("matrix_bd_63")]; + tensor var_3589 = add(x = matrix_ac_31, y = matrix_bd_63)[name = tensor("op_3589")]; + tensor _inversed_scores_61_y_0 = const()[name = tensor("_inversed_scores_61_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_61 = mul(x = var_3589, y = _inversed_scores_61_y_0)[name = tensor("_inversed_scores_61")]; + tensor scores_63 = select(a = var_41, b = _inversed_scores_61, cond = mask_11)[name = tensor("scores_63")]; + tensor var_3595 = softmax(axis = var_56, x = scores_63)[name = tensor("op_3595")]; + tensor input_821 = select(a = var_40, b = var_3595, cond = mask_11)[name = tensor("input_821")]; + tensor x_403_transpose_x_0 = const()[name = tensor("x_403_transpose_x_0"), val = tensor(false)]; + tensor x_403_transpose_y_0 = const()[name = tensor("x_403_transpose_y_0"), val = tensor(false)]; + tensor value_39 = transpose(perm = value_39_perm_0, x = v_31)[name = tensor("transpose_224")]; + tensor x_403 = matmul(transpose_x = x_403_transpose_x_0, transpose_y = x_403_transpose_y_0, x = input_821, y = value_39)[name = tensor("x_403")]; + tensor var_3599_perm_0 = const()[name = tensor("op_3599_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3600 = const()[name = tensor("op_3600"), val = tensor([1, -1, 1024])]; + tensor var_3599 = transpose(perm = var_3599_perm_0, x = x_403)[name = tensor("transpose_223")]; + tensor input_823 = reshape(shape = var_3600, x = var_3599)[name = tensor("input_823")]; + tensor input_825 = linear(bias = linear_2_bias_0, weight = module_layers_15_self_attn_linear_out_weight_quantized, x = input_823)[name = tensor("linear_142")]; + tensor input_827 = add(x = input_817, y = input_825)[name = tensor("input_827")]; + tensor x_407_axes_0 = const()[name = tensor("x_407_axes_0"), val = tensor([-1])]; + tensor x_407 = layer_norm(axes = x_407_axes_0, beta = module_layers_15_norm_conv_bias, epsilon = var_38, gamma = module_layers_15_norm_conv_weight, x = input_827)[name = tensor("x_407")]; + tensor input_829_perm_0 = const()[name = tensor("input_829_perm_0"), val = tensor([0, 2, 1])]; + tensor input_831_pad_type_0 = const()[name = tensor("input_831_pad_type_0"), val = tensor("valid")]; + tensor input_831_strides_0 = const()[name = tensor("input_831_strides_0"), val = tensor([1])]; + tensor input_831_pad_0 = const()[name = tensor("input_831_pad_0"), val = tensor([0, 0])]; + tensor input_831_dilations_0 = const()[name = tensor("input_831_dilations_0"), val = tensor([1])]; + tensor input_831_groups_0 = const()[name = tensor("input_831_groups_0"), val = tensor(1)]; + tensor input_829 = transpose(perm = input_829_perm_0, x = x_407)[name = tensor("transpose_222")]; + tensor input_831 = conv(dilations = input_831_dilations_0, groups = input_831_groups_0, pad = input_831_pad_0, pad_type = input_831_pad_type_0, strides = input_831_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_quantized, x = input_829)[name = tensor("input_831")]; + tensor x_409_split_num_splits_0 = const()[name = tensor("x_409_split_num_splits_0"), val = tensor(2)]; + tensor x_409_split_axis_0 = const()[name = tensor("x_409_split_axis_0"), val = tensor(1)]; + tensor x_409_split_0, tensor x_409_split_1 = split(axis = x_409_split_axis_0, num_splits = x_409_split_num_splits_0, x = input_831)[name = tensor("x_409_split")]; + tensor x_409_split_1_sigmoid = sigmoid(x = x_409_split_1)[name = tensor("x_409_split_1_sigmoid")]; + tensor x_409 = mul(x = x_409_split_0, y = x_409_split_1_sigmoid)[name = tensor("x_409")]; + tensor input_833 = select(a = var_40, b = x_409, cond = var_565)[name = tensor("input_833")]; + tensor new_x_63_interleave_0 = const()[name = tensor("new_x_63_interleave_0"), val = tensor(false)]; + tensor new_x_63 = concat(axis = var_56, interleave = new_x_63_interleave_0, values = (cache_63, input_833))[name = tensor("new_x_63")]; + tensor var_3638_begin_0 = const()[name = tensor("op_3638_begin_0"), val = tensor([0, 0, 2])]; + tensor var_3638_end_0 = const()[name = tensor("op_3638_end_0"), val = tensor([1, 1024, 10])]; + tensor var_3638_end_mask_0 = const()[name = tensor("op_3638_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3638 = slice_by_index(begin = var_3638_begin_0, end = var_3638_end_0, end_mask = var_3638_end_mask_0, x = new_x_63)[name = tensor("op_3638")]; + tensor x_411_pad_type_0 = const()[name = tensor("x_411_pad_type_0"), val = tensor("valid")]; + tensor x_411_groups_0 = const()[name = tensor("x_411_groups_0"), val = tensor(1024)]; + tensor x_411_strides_0 = const()[name = tensor("x_411_strides_0"), val = tensor([1])]; + tensor x_411_pad_0 = const()[name = tensor("x_411_pad_0"), val = tensor([0, 0])]; + tensor x_411_dilations_0 = const()[name = tensor("x_411_dilations_0"), val = tensor([1])]; + tensor x_411 = conv(dilations = x_411_dilations_0, groups = x_411_groups_0, pad = x_411_pad_0, pad_type = x_411_pad_type_0, strides = x_411_strides_0, weight = module_layers_15_conv_depthwise_conv_weight_quantized, x = new_x_63)[name = tensor("x_411")]; + tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; + tensor x_413_axes_0 = const()[name = tensor("x_413_axes_0"), val = tensor([-1])]; + tensor input_835 = transpose(perm = input_835_perm_0, x = x_411)[name = tensor("transpose_221")]; + tensor x_413 = layer_norm(axes = x_413_axes_0, beta = module_layers_15_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_15_conv_batch_norm_weight, x = input_835)[name = tensor("x_413")]; + tensor input_837_perm_0 = const()[name = tensor("input_837_perm_0"), val = tensor([0, 2, 1])]; + tensor input_837 = transpose(perm = input_837_perm_0, x = x_413)[name = tensor("transpose_220")]; + tensor input_839 = silu(x = input_837)[name = tensor("input_839")]; + tensor x_415_pad_type_0 = const()[name = tensor("x_415_pad_type_0"), val = tensor("valid")]; + tensor x_415_strides_0 = const()[name = tensor("x_415_strides_0"), val = tensor([1])]; + tensor x_415_pad_0 = const()[name = tensor("x_415_pad_0"), val = tensor([0, 0])]; + tensor x_415_dilations_0 = const()[name = tensor("x_415_dilations_0"), val = tensor([1])]; + tensor x_415_groups_0 = const()[name = tensor("x_415_groups_0"), val = tensor(1)]; + tensor x_415 = conv(dilations = x_415_dilations_0, groups = x_415_groups_0, pad = x_415_pad_0, pad_type = x_415_pad_type_0, strides = x_415_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_quantized, x = input_839)[name = tensor("x_415")]; + tensor input_841_perm_0 = const()[name = tensor("input_841_perm_0"), val = tensor([0, 2, 1])]; + tensor input_841 = transpose(perm = input_841_perm_0, x = x_415)[name = tensor("transpose_219")]; + tensor input_843 = add(x = input_827, y = input_841)[name = tensor("input_843")]; + tensor input_845_axes_0 = const()[name = tensor("input_845_axes_0"), val = tensor([-1])]; + tensor input_845 = layer_norm(axes = input_845_axes_0, beta = module_layers_15_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_15_norm_feed_forward2_weight, x = input_843)[name = tensor("input_845")]; + tensor input_847 = linear(bias = linear_1_bias_0, weight = module_layers_15_feed_forward2_linear1_weight_quantized, x = input_845)[name = tensor("linear_143")]; + tensor input_849 = silu(x = input_847)[name = tensor("input_849")]; + tensor input_853 = linear(bias = linear_2_bias_0, weight = module_layers_15_feed_forward2_linear2_weight_quantized, x = input_849)[name = tensor("linear_144")]; + tensor var_3679 = const()[name = tensor("op_3679"), val = tensor(0x1p-1)]; + tensor var_3680 = mul(x = input_853, y = var_3679)[name = tensor("op_3680")]; + tensor input_855 = add(x = input_843, y = var_3680)[name = tensor("input_855")]; + tensor input_857_axes_0 = const()[name = tensor("input_857_axes_0"), val = tensor([-1])]; + tensor input_857 = layer_norm(axes = input_857_axes_0, beta = module_layers_15_norm_out_bias, epsilon = var_38, gamma = module_layers_15_norm_out_weight, x = input_855)[name = tensor("input_857")]; + tensor cache_65_begin_0 = const()[name = tensor("cache_65_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_65_end_0 = const()[name = tensor("cache_65_end_0"), val = tensor([17, 1, 70, 1024])]; + tensor cache_65_end_mask_0 = const()[name = tensor("cache_65_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_65_squeeze_mask_0 = const()[name = tensor("cache_65_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_65 = slice_by_index(begin = cache_65_begin_0, end = cache_65_end_0, end_mask = cache_65_end_mask_0, squeeze_mask = cache_65_squeeze_mask_0, x = value_3)[name = tensor("cache_65")]; + tensor cache_67_begin_0 = const()[name = tensor("cache_67_begin_0"), val = tensor([16, 0, 0, 0])]; + tensor cache_67_end_0 = const()[name = tensor("cache_67_end_0"), val = tensor([17, 1, 1024, 8])]; + tensor cache_67_end_mask_0 = const()[name = tensor("cache_67_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_67_squeeze_mask_0 = const()[name = tensor("cache_67_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_67 = slice_by_index(begin = cache_67_begin_0, end = cache_67_end_0, end_mask = cache_67_end_mask_0, squeeze_mask = cache_67_squeeze_mask_0, x = value_5)[name = tensor("cache_67")]; + tensor input_859_axes_0 = const()[name = tensor("input_859_axes_0"), val = tensor([-1])]; + tensor input_859 = layer_norm(axes = input_859_axes_0, beta = module_layers_16_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_16_norm_feed_forward1_weight, x = input_857)[name = tensor("input_859")]; + tensor input_861 = linear(bias = linear_1_bias_0, weight = module_layers_16_feed_forward1_linear1_weight_quantized, x = input_859)[name = tensor("linear_145")]; + tensor input_863 = silu(x = input_861)[name = tensor("input_863")]; + tensor input_867 = linear(bias = linear_2_bias_0, weight = module_layers_16_feed_forward1_linear2_weight_quantized, x = input_863)[name = tensor("linear_146")]; + tensor var_3714 = const()[name = tensor("op_3714"), val = tensor(0x1p-1)]; + tensor var_3715 = mul(x = input_867, y = var_3714)[name = tensor("op_3715")]; + tensor input_869 = add(x = input_857, y = var_3715)[name = tensor("input_869")]; + tensor key_33_axes_0 = const()[name = tensor("key_33_axes_0"), val = tensor([-1])]; + tensor key_33 = layer_norm(axes = key_33_axes_0, beta = module_layers_16_norm_self_att_bias, epsilon = var_38, gamma = module_layers_16_norm_self_att_weight, x = input_869)[name = tensor("key_33")]; + tensor input_871_interleave_0 = const()[name = tensor("input_871_interleave_0"), val = tensor(false)]; + tensor input_871 = concat(axis = var_65, interleave = input_871_interleave_0, values = (cache_65, key_33))[name = tensor("input_871")]; + tensor var_3737_begin_0 = const()[name = tensor("op_3737_begin_0"), val = tensor([0, 2, 0])]; + tensor var_3737_end_0 = const()[name = tensor("op_3737_end_0"), val = tensor([1, 70, 1024])]; + tensor var_3737_end_mask_0 = const()[name = tensor("op_3737_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3737 = slice_by_index(begin = var_3737_begin_0, end = var_3737_end_0, end_mask = var_3737_end_mask_0, x = cache_65)[name = tensor("op_3737")]; + tensor var_3743_interleave_0 = const()[name = tensor("op_3743_interleave_0"), val = tensor(false)]; + tensor var_3743 = concat(axis = var_65, interleave = var_3743_interleave_0, values = (var_3737, key_33))[name = tensor("op_3743")]; + tensor var_3746 = linear(bias = linear_2_bias_0, weight = module_layers_16_self_attn_linear_q_weight_quantized, x = key_33)[name = tensor("linear_147")]; + tensor var_3747 = const()[name = tensor("op_3747"), val = tensor([1, -1, 8, 128])]; + tensor q_97 = reshape(shape = var_3747, x = var_3746)[name = tensor("q_97")]; + tensor var_3750 = linear(bias = linear_2_bias_0, weight = module_layers_16_self_attn_linear_k_weight_quantized, x = input_871)[name = tensor("linear_148")]; + tensor var_3751 = const()[name = tensor("op_3751"), val = tensor([1, -1, 8, 128])]; + tensor k_65 = reshape(shape = var_3751, x = var_3750)[name = tensor("k_65")]; + tensor var_3754 = linear(bias = linear_2_bias_0, weight = module_layers_16_self_attn_linear_v_weight_quantized, x = input_871)[name = tensor("linear_149")]; + tensor var_3755 = const()[name = tensor("op_3755"), val = tensor([1, -1, 8, 128])]; + tensor v_33 = reshape(shape = var_3755, x = var_3754)[name = tensor("v_33")]; + tensor value_41_perm_0 = const()[name = tensor("value_41_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_3767 = add(x = q_97, y = module_layers_16_self_attn_pos_bias_u)[name = tensor("op_3767")]; + tensor var_3769 = add(x = q_97, y = module_layers_16_self_attn_pos_bias_v)[name = tensor("op_3769")]; + tensor q_with_bias_v_33_perm_0 = const()[name = tensor("q_with_bias_v_33_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_3771_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3771_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589691456))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589838208))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589837952)))]; + tensor x_423_transpose_x_0 = const()[name = tensor("x_423_transpose_x_0"), val = tensor(false)]; + tensor x_423_transpose_y_0 = const()[name = tensor("x_423_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_33 = transpose(perm = q_with_bias_v_33_perm_0, x = var_3769)[name = tensor("transpose_218")]; + tensor x_423 = matmul(transpose_x = x_423_transpose_x_0, transpose_y = x_423_transpose_y_0, x = q_with_bias_v_33, y = op_3771_quantized)[name = tensor("x_423")]; + tensor const_287 = const()[name = tensor("const_287"), val = tensor(0x0p+0)]; + tensor x_425_pad_0 = const()[name = tensor("x_425_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_425_mode_0 = const()[name = tensor("x_425_mode_0"), val = tensor("constant")]; + tensor x_425 = pad(constant_val = const_287, mode = x_425_mode_0, pad = x_425_pad_0, x = x_423)[name = tensor("x_425")]; + tensor var_3779 = const()[name = tensor("op_3779"), val = tensor([1, 8, -1, 2])]; + tensor x_427 = reshape(shape = var_3779, x = x_425)[name = tensor("x_427")]; + tensor var_3783_begin_0 = const()[name = tensor("op_3783_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3783_end_0 = const()[name = tensor("op_3783_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_3783_end_mask_0 = const()[name = tensor("op_3783_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3783 = slice_by_index(begin = var_3783_begin_0, end = var_3783_end_0, end_mask = var_3783_end_mask_0, x = x_427)[name = tensor("op_3783")]; + tensor var_3784 = const()[name = tensor("op_3784"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_65 = reshape(shape = var_3784, x = var_3783)[name = tensor("matrix_bd_65")]; + tensor matrix_ac_33_transpose_x_0 = const()[name = tensor("matrix_ac_33_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_33_transpose_y_0 = const()[name = tensor("matrix_ac_33_transpose_y_0"), val = tensor(false)]; + tensor transpose_128_perm_0 = const()[name = tensor("transpose_128_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_129_perm_0 = const()[name = tensor("transpose_129_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_129 = transpose(perm = transpose_129_perm_0, x = k_65)[name = tensor("transpose_216")]; + tensor transpose_128 = transpose(perm = transpose_128_perm_0, x = var_3767)[name = tensor("transpose_217")]; + tensor matrix_ac_33 = matmul(transpose_x = matrix_ac_33_transpose_x_0, transpose_y = matrix_ac_33_transpose_y_0, x = transpose_128, y = transpose_129)[name = tensor("matrix_ac_33")]; + tensor matrix_bd_67_begin_0 = const()[name = tensor("matrix_bd_67_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_67_end_0 = const()[name = tensor("matrix_bd_67_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_67_end_mask_0 = const()[name = tensor("matrix_bd_67_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_67 = slice_by_index(begin = matrix_bd_67_begin_0, end = matrix_bd_67_end_0, end_mask = matrix_bd_67_end_mask_0, x = matrix_bd_65)[name = tensor("matrix_bd_67")]; + tensor var_3793 = add(x = matrix_ac_33, y = matrix_bd_67)[name = tensor("op_3793")]; + tensor _inversed_scores_65_y_0 = const()[name = tensor("_inversed_scores_65_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_65 = mul(x = var_3793, y = _inversed_scores_65_y_0)[name = tensor("_inversed_scores_65")]; + tensor scores_67 = select(a = var_41, b = _inversed_scores_65, cond = mask_11)[name = tensor("scores_67")]; + tensor var_3799 = softmax(axis = var_56, x = scores_67)[name = tensor("op_3799")]; + tensor input_873 = select(a = var_40, b = var_3799, cond = mask_11)[name = tensor("input_873")]; + tensor x_429_transpose_x_0 = const()[name = tensor("x_429_transpose_x_0"), val = tensor(false)]; + tensor x_429_transpose_y_0 = const()[name = tensor("x_429_transpose_y_0"), val = tensor(false)]; + tensor value_41 = transpose(perm = value_41_perm_0, x = v_33)[name = tensor("transpose_215")]; + tensor x_429 = matmul(transpose_x = x_429_transpose_x_0, transpose_y = x_429_transpose_y_0, x = input_873, y = value_41)[name = tensor("x_429")]; + tensor var_3803_perm_0 = const()[name = tensor("op_3803_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_3804 = const()[name = tensor("op_3804"), val = tensor([1, -1, 1024])]; + tensor var_3803 = transpose(perm = var_3803_perm_0, x = x_429)[name = tensor("transpose_214")]; + tensor input_875 = reshape(shape = var_3804, x = var_3803)[name = tensor("input_875")]; + tensor input_877 = linear(bias = linear_2_bias_0, weight = module_layers_16_self_attn_linear_out_weight_quantized, x = input_875)[name = tensor("linear_151")]; + tensor input_879 = add(x = input_869, y = input_877)[name = tensor("input_879")]; + tensor x_433_axes_0 = const()[name = tensor("x_433_axes_0"), val = tensor([-1])]; + tensor x_433 = layer_norm(axes = x_433_axes_0, beta = module_layers_16_norm_conv_bias, epsilon = var_38, gamma = module_layers_16_norm_conv_weight, x = input_879)[name = tensor("x_433")]; + tensor input_881_perm_0 = const()[name = tensor("input_881_perm_0"), val = tensor([0, 2, 1])]; + tensor input_883_pad_type_0 = const()[name = tensor("input_883_pad_type_0"), val = tensor("valid")]; + tensor input_883_strides_0 = const()[name = tensor("input_883_strides_0"), val = tensor([1])]; + tensor input_883_pad_0 = const()[name = tensor("input_883_pad_0"), val = tensor([0, 0])]; + tensor input_883_dilations_0 = const()[name = tensor("input_883_dilations_0"), val = tensor([1])]; + tensor input_883_groups_0 = const()[name = tensor("input_883_groups_0"), val = tensor(1)]; + tensor input_881 = transpose(perm = input_881_perm_0, x = x_433)[name = tensor("transpose_213")]; + tensor input_883 = conv(dilations = input_883_dilations_0, groups = input_883_groups_0, pad = input_883_pad_0, pad_type = input_883_pad_type_0, strides = input_883_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_quantized, x = input_881)[name = tensor("input_883")]; + tensor x_435_split_num_splits_0 = const()[name = tensor("x_435_split_num_splits_0"), val = tensor(2)]; + tensor x_435_split_axis_0 = const()[name = tensor("x_435_split_axis_0"), val = tensor(1)]; + tensor x_435_split_0, tensor x_435_split_1 = split(axis = x_435_split_axis_0, num_splits = x_435_split_num_splits_0, x = input_883)[name = tensor("x_435_split")]; + tensor x_435_split_1_sigmoid = sigmoid(x = x_435_split_1)[name = tensor("x_435_split_1_sigmoid")]; + tensor x_435 = mul(x = x_435_split_0, y = x_435_split_1_sigmoid)[name = tensor("x_435")]; + tensor input_885 = select(a = var_40, b = x_435, cond = var_565)[name = tensor("input_885")]; + tensor new_x_67_interleave_0 = const()[name = tensor("new_x_67_interleave_0"), val = tensor(false)]; + tensor new_x_67 = concat(axis = var_56, interleave = new_x_67_interleave_0, values = (cache_67, input_885))[name = tensor("new_x_67")]; + tensor var_3842_begin_0 = const()[name = tensor("op_3842_begin_0"), val = tensor([0, 0, 2])]; + tensor var_3842_end_0 = const()[name = tensor("op_3842_end_0"), val = tensor([1, 1024, 10])]; + tensor var_3842_end_mask_0 = const()[name = tensor("op_3842_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3842 = slice_by_index(begin = var_3842_begin_0, end = var_3842_end_0, end_mask = var_3842_end_mask_0, x = new_x_67)[name = tensor("op_3842")]; + tensor x_437_pad_type_0 = const()[name = tensor("x_437_pad_type_0"), val = tensor("valid")]; + tensor x_437_groups_0 = const()[name = tensor("x_437_groups_0"), val = tensor(1024)]; + tensor x_437_strides_0 = const()[name = tensor("x_437_strides_0"), val = tensor([1])]; + tensor x_437_pad_0 = const()[name = tensor("x_437_pad_0"), val = tensor([0, 0])]; + tensor x_437_dilations_0 = const()[name = tensor("x_437_dilations_0"), val = tensor([1])]; + tensor x_437 = conv(dilations = x_437_dilations_0, groups = x_437_groups_0, pad = x_437_pad_0, pad_type = x_437_pad_type_0, strides = x_437_strides_0, weight = module_layers_16_conv_depthwise_conv_weight_quantized, x = new_x_67)[name = tensor("x_437")]; + tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; + tensor x_439_axes_0 = const()[name = tensor("x_439_axes_0"), val = tensor([-1])]; + tensor input_887 = transpose(perm = input_887_perm_0, x = x_437)[name = tensor("transpose_212")]; + tensor x_439 = layer_norm(axes = x_439_axes_0, beta = module_layers_16_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_16_conv_batch_norm_weight, x = input_887)[name = tensor("x_439")]; + tensor input_889_perm_0 = const()[name = tensor("input_889_perm_0"), val = tensor([0, 2, 1])]; + tensor input_889 = transpose(perm = input_889_perm_0, x = x_439)[name = tensor("transpose_211")]; + tensor input_891 = silu(x = input_889)[name = tensor("input_891")]; + tensor x_441_pad_type_0 = const()[name = tensor("x_441_pad_type_0"), val = tensor("valid")]; + tensor x_441_strides_0 = const()[name = tensor("x_441_strides_0"), val = tensor([1])]; + tensor x_441_pad_0 = const()[name = tensor("x_441_pad_0"), val = tensor([0, 0])]; + tensor x_441_dilations_0 = const()[name = tensor("x_441_dilations_0"), val = tensor([1])]; + tensor x_441_groups_0 = const()[name = tensor("x_441_groups_0"), val = tensor(1)]; + tensor x_441 = conv(dilations = x_441_dilations_0, groups = x_441_groups_0, pad = x_441_pad_0, pad_type = x_441_pad_type_0, strides = x_441_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_quantized, x = input_891)[name = tensor("x_441")]; + tensor input_893_perm_0 = const()[name = tensor("input_893_perm_0"), val = tensor([0, 2, 1])]; + tensor input_893 = transpose(perm = input_893_perm_0, x = x_441)[name = tensor("transpose_210")]; + tensor input_895 = add(x = input_879, y = input_893)[name = tensor("input_895")]; + tensor input_897_axes_0 = const()[name = tensor("input_897_axes_0"), val = tensor([-1])]; + tensor input_897 = layer_norm(axes = input_897_axes_0, beta = module_layers_16_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_16_norm_feed_forward2_weight, x = input_895)[name = tensor("input_897")]; + tensor input_899 = linear(bias = linear_1_bias_0, weight = module_layers_16_feed_forward2_linear1_weight_quantized, x = input_897)[name = tensor("linear_152")]; + tensor input_901 = silu(x = input_899)[name = tensor("input_901")]; + tensor input_905 = linear(bias = linear_2_bias_0, weight = module_layers_16_feed_forward2_linear2_weight_quantized, x = input_901)[name = tensor("linear_153")]; + tensor var_3883 = const()[name = tensor("op_3883"), val = tensor(0x1p-1)]; + tensor var_3884 = mul(x = input_905, y = var_3883)[name = tensor("op_3884")]; + tensor input_907 = add(x = input_895, y = var_3884)[name = tensor("input_907")]; + tensor input_909_axes_0 = const()[name = tensor("input_909_axes_0"), val = tensor([-1])]; + tensor input_909 = layer_norm(axes = input_909_axes_0, beta = module_layers_16_norm_out_bias, epsilon = var_38, gamma = module_layers_16_norm_out_weight, x = input_907)[name = tensor("input_909")]; + tensor cache_69_begin_0 = const()[name = tensor("cache_69_begin_0"), val = tensor([17, 0, 0, 0])]; + tensor cache_69_end_0 = const()[name = tensor("cache_69_end_0"), val = tensor([18, 1, 70, 1024])]; + tensor cache_69_end_mask_0 = const()[name = tensor("cache_69_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_69_squeeze_mask_0 = const()[name = tensor("cache_69_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_69 = slice_by_index(begin = cache_69_begin_0, end = cache_69_end_0, end_mask = cache_69_end_mask_0, squeeze_mask = cache_69_squeeze_mask_0, x = value_3)[name = tensor("cache_69")]; + tensor cache_71_begin_0 = const()[name = tensor("cache_71_begin_0"), val = tensor([17, 0, 0, 0])]; + tensor cache_71_end_0 = const()[name = tensor("cache_71_end_0"), val = tensor([18, 1, 1024, 8])]; + tensor cache_71_end_mask_0 = const()[name = tensor("cache_71_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_71_squeeze_mask_0 = const()[name = tensor("cache_71_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_71 = slice_by_index(begin = cache_71_begin_0, end = cache_71_end_0, end_mask = cache_71_end_mask_0, squeeze_mask = cache_71_squeeze_mask_0, x = value_5)[name = tensor("cache_71")]; + tensor input_911_axes_0 = const()[name = tensor("input_911_axes_0"), val = tensor([-1])]; + tensor input_911 = layer_norm(axes = input_911_axes_0, beta = module_layers_17_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_17_norm_feed_forward1_weight, x = input_909)[name = tensor("input_911")]; + tensor input_913 = linear(bias = linear_1_bias_0, weight = module_layers_17_feed_forward1_linear1_weight_quantized, x = input_911)[name = tensor("linear_154")]; + tensor input_915 = silu(x = input_913)[name = tensor("input_915")]; + tensor input_919 = linear(bias = linear_2_bias_0, weight = module_layers_17_feed_forward1_linear2_weight_quantized, x = input_915)[name = tensor("linear_155")]; + tensor var_3918 = const()[name = tensor("op_3918"), val = tensor(0x1p-1)]; + tensor var_3919 = mul(x = input_919, y = var_3918)[name = tensor("op_3919")]; + tensor input_921 = add(x = input_909, y = var_3919)[name = tensor("input_921")]; + tensor key_35_axes_0 = const()[name = tensor("key_35_axes_0"), val = tensor([-1])]; + tensor key_35 = layer_norm(axes = key_35_axes_0, beta = module_layers_17_norm_self_att_bias, epsilon = var_38, gamma = module_layers_17_norm_self_att_weight, x = input_921)[name = tensor("key_35")]; + tensor input_923_interleave_0 = const()[name = tensor("input_923_interleave_0"), val = tensor(false)]; + tensor input_923 = concat(axis = var_65, interleave = input_923_interleave_0, values = (cache_69, key_35))[name = tensor("input_923")]; + tensor var_3941_begin_0 = const()[name = tensor("op_3941_begin_0"), val = tensor([0, 2, 0])]; + tensor var_3941_end_0 = const()[name = tensor("op_3941_end_0"), val = tensor([1, 70, 1024])]; + tensor var_3941_end_mask_0 = const()[name = tensor("op_3941_end_mask_0"), val = tensor([true, true, true])]; + tensor var_3941 = slice_by_index(begin = var_3941_begin_0, end = var_3941_end_0, end_mask = var_3941_end_mask_0, x = cache_69)[name = tensor("op_3941")]; + tensor var_3947_interleave_0 = const()[name = tensor("op_3947_interleave_0"), val = tensor(false)]; + tensor var_3947 = concat(axis = var_65, interleave = var_3947_interleave_0, values = (var_3941, key_35))[name = tensor("op_3947")]; + tensor var_3950 = linear(bias = linear_2_bias_0, weight = module_layers_17_self_attn_linear_q_weight_quantized, x = key_35)[name = tensor("linear_156")]; + tensor var_3951 = const()[name = tensor("op_3951"), val = tensor([1, -1, 8, 128])]; + tensor q_103 = reshape(shape = var_3951, x = var_3950)[name = tensor("q_103")]; + tensor var_3954 = linear(bias = linear_2_bias_0, weight = module_layers_17_self_attn_linear_k_weight_quantized, x = input_923)[name = tensor("linear_157")]; + tensor var_3955 = const()[name = tensor("op_3955"), val = tensor([1, -1, 8, 128])]; + tensor k_69 = reshape(shape = var_3955, x = var_3954)[name = tensor("k_69")]; + tensor var_3958 = linear(bias = linear_2_bias_0, weight = module_layers_17_self_attn_linear_v_weight_quantized, x = input_923)[name = tensor("linear_158")]; + tensor var_3959 = const()[name = tensor("op_3959"), val = tensor([1, -1, 8, 128])]; + tensor v_35 = reshape(shape = var_3959, x = var_3958)[name = tensor("v_35")]; + tensor value_43_perm_0 = const()[name = tensor("value_43_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_3971 = add(x = q_103, y = module_layers_17_self_attn_pos_bias_u)[name = tensor("op_3971")]; + tensor var_3973 = add(x = q_103, y = module_layers_17_self_attn_pos_bias_v)[name = tensor("op_3973")]; + tensor q_with_bias_v_35_perm_0 = const()[name = tensor("q_with_bias_v_35_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_3975_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_3975_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589838848))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589985600))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589985344)))]; + tensor x_449_transpose_x_0 = const()[name = tensor("x_449_transpose_x_0"), val = tensor(false)]; + tensor x_449_transpose_y_0 = const()[name = tensor("x_449_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_35 = transpose(perm = q_with_bias_v_35_perm_0, x = var_3973)[name = tensor("transpose_209")]; + tensor x_449 = matmul(transpose_x = x_449_transpose_x_0, transpose_y = x_449_transpose_y_0, x = q_with_bias_v_35, y = op_3975_quantized)[name = tensor("x_449")]; + tensor const_300 = const()[name = tensor("const_300"), val = tensor(0x0p+0)]; + tensor x_451_pad_0 = const()[name = tensor("x_451_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_451_mode_0 = const()[name = tensor("x_451_mode_0"), val = tensor("constant")]; + tensor x_451 = pad(constant_val = const_300, mode = x_451_mode_0, pad = x_451_pad_0, x = x_449)[name = tensor("x_451")]; + tensor var_3983 = const()[name = tensor("op_3983"), val = tensor([1, 8, -1, 2])]; + tensor x_453 = reshape(shape = var_3983, x = x_451)[name = tensor("x_453")]; + tensor var_3987_begin_0 = const()[name = tensor("op_3987_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_3987_end_0 = const()[name = tensor("op_3987_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_3987_end_mask_0 = const()[name = tensor("op_3987_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_3987 = slice_by_index(begin = var_3987_begin_0, end = var_3987_end_0, end_mask = var_3987_end_mask_0, x = x_453)[name = tensor("op_3987")]; + tensor var_3988 = const()[name = tensor("op_3988"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_69 = reshape(shape = var_3988, x = var_3987)[name = tensor("matrix_bd_69")]; + tensor matrix_ac_35_transpose_x_0 = const()[name = tensor("matrix_ac_35_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_35_transpose_y_0 = const()[name = tensor("matrix_ac_35_transpose_y_0"), val = tensor(false)]; + tensor transpose_130_perm_0 = const()[name = tensor("transpose_130_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_131_perm_0 = const()[name = tensor("transpose_131_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_131 = transpose(perm = transpose_131_perm_0, x = k_69)[name = tensor("transpose_207")]; + tensor transpose_130 = transpose(perm = transpose_130_perm_0, x = var_3971)[name = tensor("transpose_208")]; + tensor matrix_ac_35 = matmul(transpose_x = matrix_ac_35_transpose_x_0, transpose_y = matrix_ac_35_transpose_y_0, x = transpose_130, y = transpose_131)[name = tensor("matrix_ac_35")]; + tensor matrix_bd_71_begin_0 = const()[name = tensor("matrix_bd_71_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_71_end_0 = const()[name = tensor("matrix_bd_71_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_71_end_mask_0 = const()[name = tensor("matrix_bd_71_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_71 = slice_by_index(begin = matrix_bd_71_begin_0, end = matrix_bd_71_end_0, end_mask = matrix_bd_71_end_mask_0, x = matrix_bd_69)[name = tensor("matrix_bd_71")]; + tensor var_3997 = add(x = matrix_ac_35, y = matrix_bd_71)[name = tensor("op_3997")]; + tensor _inversed_scores_69_y_0 = const()[name = tensor("_inversed_scores_69_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_69 = mul(x = var_3997, y = _inversed_scores_69_y_0)[name = tensor("_inversed_scores_69")]; + tensor scores_71 = select(a = var_41, b = _inversed_scores_69, cond = mask_11)[name = tensor("scores_71")]; + tensor var_4003 = softmax(axis = var_56, x = scores_71)[name = tensor("op_4003")]; + tensor input_925 = select(a = var_40, b = var_4003, cond = mask_11)[name = tensor("input_925")]; + tensor x_455_transpose_x_0 = const()[name = tensor("x_455_transpose_x_0"), val = tensor(false)]; + tensor x_455_transpose_y_0 = const()[name = tensor("x_455_transpose_y_0"), val = tensor(false)]; + tensor value_43 = transpose(perm = value_43_perm_0, x = v_35)[name = tensor("transpose_206")]; + tensor x_455 = matmul(transpose_x = x_455_transpose_x_0, transpose_y = x_455_transpose_y_0, x = input_925, y = value_43)[name = tensor("x_455")]; + tensor var_4007_perm_0 = const()[name = tensor("op_4007_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_4008 = const()[name = tensor("op_4008"), val = tensor([1, -1, 1024])]; + tensor var_4007 = transpose(perm = var_4007_perm_0, x = x_455)[name = tensor("transpose_205")]; + tensor input_927 = reshape(shape = var_4008, x = var_4007)[name = tensor("input_927")]; + tensor input_929 = linear(bias = linear_2_bias_0, weight = module_layers_17_self_attn_linear_out_weight_quantized, x = input_927)[name = tensor("linear_160")]; + tensor input_931 = add(x = input_921, y = input_929)[name = tensor("input_931")]; + tensor x_459_axes_0 = const()[name = tensor("x_459_axes_0"), val = tensor([-1])]; + tensor x_459 = layer_norm(axes = x_459_axes_0, beta = module_layers_17_norm_conv_bias, epsilon = var_38, gamma = module_layers_17_norm_conv_weight, x = input_931)[name = tensor("x_459")]; + tensor input_933_perm_0 = const()[name = tensor("input_933_perm_0"), val = tensor([0, 2, 1])]; + tensor input_935_pad_type_0 = const()[name = tensor("input_935_pad_type_0"), val = tensor("valid")]; + tensor input_935_strides_0 = const()[name = tensor("input_935_strides_0"), val = tensor([1])]; + tensor input_935_pad_0 = const()[name = tensor("input_935_pad_0"), val = tensor([0, 0])]; + tensor input_935_dilations_0 = const()[name = tensor("input_935_dilations_0"), val = tensor([1])]; + tensor input_935_groups_0 = const()[name = tensor("input_935_groups_0"), val = tensor(1)]; + tensor input_933 = transpose(perm = input_933_perm_0, x = x_459)[name = tensor("transpose_204")]; + tensor input_935 = conv(dilations = input_935_dilations_0, groups = input_935_groups_0, pad = input_935_pad_0, pad_type = input_935_pad_type_0, strides = input_935_strides_0, weight = module_layers_17_conv_pointwise_conv1_weight_quantized, x = input_933)[name = tensor("input_935")]; + tensor x_461_split_num_splits_0 = const()[name = tensor("x_461_split_num_splits_0"), val = tensor(2)]; + tensor x_461_split_axis_0 = const()[name = tensor("x_461_split_axis_0"), val = tensor(1)]; + tensor x_461_split_0, tensor x_461_split_1 = split(axis = x_461_split_axis_0, num_splits = x_461_split_num_splits_0, x = input_935)[name = tensor("x_461_split")]; + tensor x_461_split_1_sigmoid = sigmoid(x = x_461_split_1)[name = tensor("x_461_split_1_sigmoid")]; + tensor x_461 = mul(x = x_461_split_0, y = x_461_split_1_sigmoid)[name = tensor("x_461")]; + tensor input_937 = select(a = var_40, b = x_461, cond = var_565)[name = tensor("input_937")]; + tensor new_x_71_interleave_0 = const()[name = tensor("new_x_71_interleave_0"), val = tensor(false)]; + tensor new_x_71 = concat(axis = var_56, interleave = new_x_71_interleave_0, values = (cache_71, input_937))[name = tensor("new_x_71")]; + tensor var_4046_begin_0 = const()[name = tensor("op_4046_begin_0"), val = tensor([0, 0, 2])]; + tensor var_4046_end_0 = const()[name = tensor("op_4046_end_0"), val = tensor([1, 1024, 10])]; + tensor var_4046_end_mask_0 = const()[name = tensor("op_4046_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4046 = slice_by_index(begin = var_4046_begin_0, end = var_4046_end_0, end_mask = var_4046_end_mask_0, x = new_x_71)[name = tensor("op_4046")]; + tensor x_463_pad_type_0 = const()[name = tensor("x_463_pad_type_0"), val = tensor("valid")]; + tensor x_463_groups_0 = const()[name = tensor("x_463_groups_0"), val = tensor(1024)]; + tensor x_463_strides_0 = const()[name = tensor("x_463_strides_0"), val = tensor([1])]; + tensor x_463_pad_0 = const()[name = tensor("x_463_pad_0"), val = tensor([0, 0])]; + tensor x_463_dilations_0 = const()[name = tensor("x_463_dilations_0"), val = tensor([1])]; + tensor x_463 = conv(dilations = x_463_dilations_0, groups = x_463_groups_0, pad = x_463_pad_0, pad_type = x_463_pad_type_0, strides = x_463_strides_0, weight = module_layers_17_conv_depthwise_conv_weight_quantized, x = new_x_71)[name = tensor("x_463")]; + tensor input_939_perm_0 = const()[name = tensor("input_939_perm_0"), val = tensor([0, 2, 1])]; + tensor x_465_axes_0 = const()[name = tensor("x_465_axes_0"), val = tensor([-1])]; + tensor input_939 = transpose(perm = input_939_perm_0, x = x_463)[name = tensor("transpose_203")]; + tensor x_465 = layer_norm(axes = x_465_axes_0, beta = module_layers_17_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_17_conv_batch_norm_weight, x = input_939)[name = tensor("x_465")]; + tensor input_941_perm_0 = const()[name = tensor("input_941_perm_0"), val = tensor([0, 2, 1])]; + tensor input_941 = transpose(perm = input_941_perm_0, x = x_465)[name = tensor("transpose_202")]; + tensor input_943 = silu(x = input_941)[name = tensor("input_943")]; + tensor x_467_pad_type_0 = const()[name = tensor("x_467_pad_type_0"), val = tensor("valid")]; + tensor x_467_strides_0 = const()[name = tensor("x_467_strides_0"), val = tensor([1])]; + tensor x_467_pad_0 = const()[name = tensor("x_467_pad_0"), val = tensor([0, 0])]; + tensor x_467_dilations_0 = const()[name = tensor("x_467_dilations_0"), val = tensor([1])]; + tensor x_467_groups_0 = const()[name = tensor("x_467_groups_0"), val = tensor(1)]; + tensor x_467 = conv(dilations = x_467_dilations_0, groups = x_467_groups_0, pad = x_467_pad_0, pad_type = x_467_pad_type_0, strides = x_467_strides_0, weight = module_layers_17_conv_pointwise_conv2_weight_quantized, x = input_943)[name = tensor("x_467")]; + tensor input_945_perm_0 = const()[name = tensor("input_945_perm_0"), val = tensor([0, 2, 1])]; + tensor input_945 = transpose(perm = input_945_perm_0, x = x_467)[name = tensor("transpose_201")]; + tensor input_947 = add(x = input_931, y = input_945)[name = tensor("input_947")]; + tensor input_949_axes_0 = const()[name = tensor("input_949_axes_0"), val = tensor([-1])]; + tensor input_949 = layer_norm(axes = input_949_axes_0, beta = module_layers_17_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_17_norm_feed_forward2_weight, x = input_947)[name = tensor("input_949")]; + tensor input_951 = linear(bias = linear_1_bias_0, weight = module_layers_17_feed_forward2_linear1_weight_quantized, x = input_949)[name = tensor("linear_161")]; + tensor input_953 = silu(x = input_951)[name = tensor("input_953")]; + tensor input_957 = linear(bias = linear_2_bias_0, weight = module_layers_17_feed_forward2_linear2_weight_quantized, x = input_953)[name = tensor("linear_162")]; + tensor var_4087 = const()[name = tensor("op_4087"), val = tensor(0x1p-1)]; + tensor var_4088 = mul(x = input_957, y = var_4087)[name = tensor("op_4088")]; + tensor input_959 = add(x = input_947, y = var_4088)[name = tensor("input_959")]; + tensor input_961_axes_0 = const()[name = tensor("input_961_axes_0"), val = tensor([-1])]; + tensor input_961 = layer_norm(axes = input_961_axes_0, beta = module_layers_17_norm_out_bias, epsilon = var_38, gamma = module_layers_17_norm_out_weight, x = input_959)[name = tensor("input_961")]; + tensor cache_73_begin_0 = const()[name = tensor("cache_73_begin_0"), val = tensor([18, 0, 0, 0])]; + tensor cache_73_end_0 = const()[name = tensor("cache_73_end_0"), val = tensor([19, 1, 70, 1024])]; + tensor cache_73_end_mask_0 = const()[name = tensor("cache_73_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_73_squeeze_mask_0 = const()[name = tensor("cache_73_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_73 = slice_by_index(begin = cache_73_begin_0, end = cache_73_end_0, end_mask = cache_73_end_mask_0, squeeze_mask = cache_73_squeeze_mask_0, x = value_3)[name = tensor("cache_73")]; + tensor cache_75_begin_0 = const()[name = tensor("cache_75_begin_0"), val = tensor([18, 0, 0, 0])]; + tensor cache_75_end_0 = const()[name = tensor("cache_75_end_0"), val = tensor([19, 1, 1024, 8])]; + tensor cache_75_end_mask_0 = const()[name = tensor("cache_75_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_75_squeeze_mask_0 = const()[name = tensor("cache_75_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_75 = slice_by_index(begin = cache_75_begin_0, end = cache_75_end_0, end_mask = cache_75_end_mask_0, squeeze_mask = cache_75_squeeze_mask_0, x = value_5)[name = tensor("cache_75")]; + tensor input_963_axes_0 = const()[name = tensor("input_963_axes_0"), val = tensor([-1])]; + tensor input_963 = layer_norm(axes = input_963_axes_0, beta = module_layers_18_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_18_norm_feed_forward1_weight, x = input_961)[name = tensor("input_963")]; + tensor input_965 = linear(bias = linear_1_bias_0, weight = module_layers_18_feed_forward1_linear1_weight_quantized, x = input_963)[name = tensor("linear_163")]; + tensor input_967 = silu(x = input_965)[name = tensor("input_967")]; + tensor input_971 = linear(bias = linear_2_bias_0, weight = module_layers_18_feed_forward1_linear2_weight_quantized, x = input_967)[name = tensor("linear_164")]; + tensor var_4122 = const()[name = tensor("op_4122"), val = tensor(0x1p-1)]; + tensor var_4123 = mul(x = input_971, y = var_4122)[name = tensor("op_4123")]; + tensor input_973 = add(x = input_961, y = var_4123)[name = tensor("input_973")]; + tensor key_37_axes_0 = const()[name = tensor("key_37_axes_0"), val = tensor([-1])]; + tensor key_37 = layer_norm(axes = key_37_axes_0, beta = module_layers_18_norm_self_att_bias, epsilon = var_38, gamma = module_layers_18_norm_self_att_weight, x = input_973)[name = tensor("key_37")]; + tensor input_975_interleave_0 = const()[name = tensor("input_975_interleave_0"), val = tensor(false)]; + tensor input_975 = concat(axis = var_65, interleave = input_975_interleave_0, values = (cache_73, key_37))[name = tensor("input_975")]; + tensor var_4145_begin_0 = const()[name = tensor("op_4145_begin_0"), val = tensor([0, 2, 0])]; + tensor var_4145_end_0 = const()[name = tensor("op_4145_end_0"), val = tensor([1, 70, 1024])]; + tensor var_4145_end_mask_0 = const()[name = tensor("op_4145_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4145 = slice_by_index(begin = var_4145_begin_0, end = var_4145_end_0, end_mask = var_4145_end_mask_0, x = cache_73)[name = tensor("op_4145")]; + tensor var_4151_interleave_0 = const()[name = tensor("op_4151_interleave_0"), val = tensor(false)]; + tensor var_4151 = concat(axis = var_65, interleave = var_4151_interleave_0, values = (var_4145, key_37))[name = tensor("op_4151")]; + tensor var_4154 = linear(bias = linear_2_bias_0, weight = module_layers_18_self_attn_linear_q_weight_quantized, x = key_37)[name = tensor("linear_165")]; + tensor var_4155 = const()[name = tensor("op_4155"), val = tensor([1, -1, 8, 128])]; + tensor q_109 = reshape(shape = var_4155, x = var_4154)[name = tensor("q_109")]; + tensor var_4158 = linear(bias = linear_2_bias_0, weight = module_layers_18_self_attn_linear_k_weight_quantized, x = input_975)[name = tensor("linear_166")]; + tensor var_4159 = const()[name = tensor("op_4159"), val = tensor([1, -1, 8, 128])]; + tensor k_73 = reshape(shape = var_4159, x = var_4158)[name = tensor("k_73")]; + tensor var_4162 = linear(bias = linear_2_bias_0, weight = module_layers_18_self_attn_linear_v_weight_quantized, x = input_975)[name = tensor("linear_167")]; + tensor var_4163 = const()[name = tensor("op_4163"), val = tensor([1, -1, 8, 128])]; + tensor v_37 = reshape(shape = var_4163, x = var_4162)[name = tensor("v_37")]; + tensor value_45_perm_0 = const()[name = tensor("value_45_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_4175 = add(x = q_109, y = module_layers_18_self_attn_pos_bias_u)[name = tensor("op_4175")]; + tensor var_4177 = add(x = q_109, y = module_layers_18_self_attn_pos_bias_v)[name = tensor("op_4177")]; + tensor q_with_bias_v_37_perm_0 = const()[name = tensor("q_with_bias_v_37_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_4179_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4179_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(589986240))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590132992))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590132736)))]; + tensor x_475_transpose_x_0 = const()[name = tensor("x_475_transpose_x_0"), val = tensor(false)]; + tensor x_475_transpose_y_0 = const()[name = tensor("x_475_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_37 = transpose(perm = q_with_bias_v_37_perm_0, x = var_4177)[name = tensor("transpose_200")]; + tensor x_475 = matmul(transpose_x = x_475_transpose_x_0, transpose_y = x_475_transpose_y_0, x = q_with_bias_v_37, y = op_4179_quantized)[name = tensor("x_475")]; + tensor const_313 = const()[name = tensor("const_313"), val = tensor(0x0p+0)]; + tensor x_477_pad_0 = const()[name = tensor("x_477_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_477_mode_0 = const()[name = tensor("x_477_mode_0"), val = tensor("constant")]; + tensor x_477 = pad(constant_val = const_313, mode = x_477_mode_0, pad = x_477_pad_0, x = x_475)[name = tensor("x_477")]; + tensor var_4187 = const()[name = tensor("op_4187"), val = tensor([1, 8, -1, 2])]; + tensor x_479 = reshape(shape = var_4187, x = x_477)[name = tensor("x_479")]; + tensor var_4191_begin_0 = const()[name = tensor("op_4191_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_4191_end_0 = const()[name = tensor("op_4191_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_4191_end_mask_0 = const()[name = tensor("op_4191_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_4191 = slice_by_index(begin = var_4191_begin_0, end = var_4191_end_0, end_mask = var_4191_end_mask_0, x = x_479)[name = tensor("op_4191")]; + tensor var_4192 = const()[name = tensor("op_4192"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_73 = reshape(shape = var_4192, x = var_4191)[name = tensor("matrix_bd_73")]; + tensor matrix_ac_37_transpose_x_0 = const()[name = tensor("matrix_ac_37_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_37_transpose_y_0 = const()[name = tensor("matrix_ac_37_transpose_y_0"), val = tensor(false)]; + tensor transpose_132_perm_0 = const()[name = tensor("transpose_132_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_133_perm_0 = const()[name = tensor("transpose_133_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_133 = transpose(perm = transpose_133_perm_0, x = k_73)[name = tensor("transpose_198")]; + tensor transpose_132 = transpose(perm = transpose_132_perm_0, x = var_4175)[name = tensor("transpose_199")]; + tensor matrix_ac_37 = matmul(transpose_x = matrix_ac_37_transpose_x_0, transpose_y = matrix_ac_37_transpose_y_0, x = transpose_132, y = transpose_133)[name = tensor("matrix_ac_37")]; + tensor matrix_bd_75_begin_0 = const()[name = tensor("matrix_bd_75_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_75_end_0 = const()[name = tensor("matrix_bd_75_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_75_end_mask_0 = const()[name = tensor("matrix_bd_75_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_75 = slice_by_index(begin = matrix_bd_75_begin_0, end = matrix_bd_75_end_0, end_mask = matrix_bd_75_end_mask_0, x = matrix_bd_73)[name = tensor("matrix_bd_75")]; + tensor var_4201 = add(x = matrix_ac_37, y = matrix_bd_75)[name = tensor("op_4201")]; + tensor _inversed_scores_73_y_0 = const()[name = tensor("_inversed_scores_73_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_73 = mul(x = var_4201, y = _inversed_scores_73_y_0)[name = tensor("_inversed_scores_73")]; + tensor scores_75 = select(a = var_41, b = _inversed_scores_73, cond = mask_11)[name = tensor("scores_75")]; + tensor var_4207 = softmax(axis = var_56, x = scores_75)[name = tensor("op_4207")]; + tensor input_977 = select(a = var_40, b = var_4207, cond = mask_11)[name = tensor("input_977")]; + tensor x_481_transpose_x_0 = const()[name = tensor("x_481_transpose_x_0"), val = tensor(false)]; + tensor x_481_transpose_y_0 = const()[name = tensor("x_481_transpose_y_0"), val = tensor(false)]; + tensor value_45 = transpose(perm = value_45_perm_0, x = v_37)[name = tensor("transpose_197")]; + tensor x_481 = matmul(transpose_x = x_481_transpose_x_0, transpose_y = x_481_transpose_y_0, x = input_977, y = value_45)[name = tensor("x_481")]; + tensor var_4211_perm_0 = const()[name = tensor("op_4211_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_4212 = const()[name = tensor("op_4212"), val = tensor([1, -1, 1024])]; + tensor var_4211 = transpose(perm = var_4211_perm_0, x = x_481)[name = tensor("transpose_196")]; + tensor input_979 = reshape(shape = var_4212, x = var_4211)[name = tensor("input_979")]; + tensor input_981 = linear(bias = linear_2_bias_0, weight = module_layers_18_self_attn_linear_out_weight_quantized, x = input_979)[name = tensor("linear_169")]; + tensor input_983 = add(x = input_973, y = input_981)[name = tensor("input_983")]; + tensor x_485_axes_0 = const()[name = tensor("x_485_axes_0"), val = tensor([-1])]; + tensor x_485 = layer_norm(axes = x_485_axes_0, beta = module_layers_18_norm_conv_bias, epsilon = var_38, gamma = module_layers_18_norm_conv_weight, x = input_983)[name = tensor("x_485")]; + tensor input_985_perm_0 = const()[name = tensor("input_985_perm_0"), val = tensor([0, 2, 1])]; + tensor input_987_pad_type_0 = const()[name = tensor("input_987_pad_type_0"), val = tensor("valid")]; + tensor input_987_strides_0 = const()[name = tensor("input_987_strides_0"), val = tensor([1])]; + tensor input_987_pad_0 = const()[name = tensor("input_987_pad_0"), val = tensor([0, 0])]; + tensor input_987_dilations_0 = const()[name = tensor("input_987_dilations_0"), val = tensor([1])]; + tensor input_987_groups_0 = const()[name = tensor("input_987_groups_0"), val = tensor(1)]; + tensor input_985 = transpose(perm = input_985_perm_0, x = x_485)[name = tensor("transpose_195")]; + tensor input_987 = conv(dilations = input_987_dilations_0, groups = input_987_groups_0, pad = input_987_pad_0, pad_type = input_987_pad_type_0, strides = input_987_strides_0, weight = module_layers_18_conv_pointwise_conv1_weight_quantized, x = input_985)[name = tensor("input_987")]; + tensor x_487_split_num_splits_0 = const()[name = tensor("x_487_split_num_splits_0"), val = tensor(2)]; + tensor x_487_split_axis_0 = const()[name = tensor("x_487_split_axis_0"), val = tensor(1)]; + tensor x_487_split_0, tensor x_487_split_1 = split(axis = x_487_split_axis_0, num_splits = x_487_split_num_splits_0, x = input_987)[name = tensor("x_487_split")]; + tensor x_487_split_1_sigmoid = sigmoid(x = x_487_split_1)[name = tensor("x_487_split_1_sigmoid")]; + tensor x_487 = mul(x = x_487_split_0, y = x_487_split_1_sigmoid)[name = tensor("x_487")]; + tensor input_989 = select(a = var_40, b = x_487, cond = var_565)[name = tensor("input_989")]; + tensor new_x_75_interleave_0 = const()[name = tensor("new_x_75_interleave_0"), val = tensor(false)]; + tensor new_x_75 = concat(axis = var_56, interleave = new_x_75_interleave_0, values = (cache_75, input_989))[name = tensor("new_x_75")]; + tensor var_4250_begin_0 = const()[name = tensor("op_4250_begin_0"), val = tensor([0, 0, 2])]; + tensor var_4250_end_0 = const()[name = tensor("op_4250_end_0"), val = tensor([1, 1024, 10])]; + tensor var_4250_end_mask_0 = const()[name = tensor("op_4250_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4250 = slice_by_index(begin = var_4250_begin_0, end = var_4250_end_0, end_mask = var_4250_end_mask_0, x = new_x_75)[name = tensor("op_4250")]; + tensor x_489_pad_type_0 = const()[name = tensor("x_489_pad_type_0"), val = tensor("valid")]; + tensor x_489_groups_0 = const()[name = tensor("x_489_groups_0"), val = tensor(1024)]; + tensor x_489_strides_0 = const()[name = tensor("x_489_strides_0"), val = tensor([1])]; + tensor x_489_pad_0 = const()[name = tensor("x_489_pad_0"), val = tensor([0, 0])]; + tensor x_489_dilations_0 = const()[name = tensor("x_489_dilations_0"), val = tensor([1])]; + tensor x_489 = conv(dilations = x_489_dilations_0, groups = x_489_groups_0, pad = x_489_pad_0, pad_type = x_489_pad_type_0, strides = x_489_strides_0, weight = module_layers_18_conv_depthwise_conv_weight_quantized, x = new_x_75)[name = tensor("x_489")]; + tensor input_991_perm_0 = const()[name = tensor("input_991_perm_0"), val = tensor([0, 2, 1])]; + tensor x_491_axes_0 = const()[name = tensor("x_491_axes_0"), val = tensor([-1])]; + tensor input_991 = transpose(perm = input_991_perm_0, x = x_489)[name = tensor("transpose_194")]; + tensor x_491 = layer_norm(axes = x_491_axes_0, beta = module_layers_18_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_18_conv_batch_norm_weight, x = input_991)[name = tensor("x_491")]; + tensor input_993_perm_0 = const()[name = tensor("input_993_perm_0"), val = tensor([0, 2, 1])]; + tensor input_993 = transpose(perm = input_993_perm_0, x = x_491)[name = tensor("transpose_193")]; + tensor input_995 = silu(x = input_993)[name = tensor("input_995")]; + tensor x_493_pad_type_0 = const()[name = tensor("x_493_pad_type_0"), val = tensor("valid")]; + tensor x_493_strides_0 = const()[name = tensor("x_493_strides_0"), val = tensor([1])]; + tensor x_493_pad_0 = const()[name = tensor("x_493_pad_0"), val = tensor([0, 0])]; + tensor x_493_dilations_0 = const()[name = tensor("x_493_dilations_0"), val = tensor([1])]; + tensor x_493_groups_0 = const()[name = tensor("x_493_groups_0"), val = tensor(1)]; + tensor x_493 = conv(dilations = x_493_dilations_0, groups = x_493_groups_0, pad = x_493_pad_0, pad_type = x_493_pad_type_0, strides = x_493_strides_0, weight = module_layers_18_conv_pointwise_conv2_weight_quantized, x = input_995)[name = tensor("x_493")]; + tensor input_997_perm_0 = const()[name = tensor("input_997_perm_0"), val = tensor([0, 2, 1])]; + tensor input_997 = transpose(perm = input_997_perm_0, x = x_493)[name = tensor("transpose_192")]; + tensor input_999 = add(x = input_983, y = input_997)[name = tensor("input_999")]; + tensor input_1001_axes_0 = const()[name = tensor("input_1001_axes_0"), val = tensor([-1])]; + tensor input_1001 = layer_norm(axes = input_1001_axes_0, beta = module_layers_18_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_18_norm_feed_forward2_weight, x = input_999)[name = tensor("input_1001")]; + tensor input_1003 = linear(bias = linear_1_bias_0, weight = module_layers_18_feed_forward2_linear1_weight_quantized, x = input_1001)[name = tensor("linear_170")]; + tensor input_1005 = silu(x = input_1003)[name = tensor("input_1005")]; + tensor input_1009 = linear(bias = linear_2_bias_0, weight = module_layers_18_feed_forward2_linear2_weight_quantized, x = input_1005)[name = tensor("linear_171")]; + tensor var_4291 = const()[name = tensor("op_4291"), val = tensor(0x1p-1)]; + tensor var_4292 = mul(x = input_1009, y = var_4291)[name = tensor("op_4292")]; + tensor input_1011 = add(x = input_999, y = var_4292)[name = tensor("input_1011")]; + tensor input_1013_axes_0 = const()[name = tensor("input_1013_axes_0"), val = tensor([-1])]; + tensor input_1013 = layer_norm(axes = input_1013_axes_0, beta = module_layers_18_norm_out_bias, epsilon = var_38, gamma = module_layers_18_norm_out_weight, x = input_1011)[name = tensor("input_1013")]; + tensor cache_77_begin_0 = const()[name = tensor("cache_77_begin_0"), val = tensor([19, 0, 0, 0])]; + tensor cache_77_end_0 = const()[name = tensor("cache_77_end_0"), val = tensor([20, 1, 70, 1024])]; + tensor cache_77_end_mask_0 = const()[name = tensor("cache_77_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_77_squeeze_mask_0 = const()[name = tensor("cache_77_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_77 = slice_by_index(begin = cache_77_begin_0, end = cache_77_end_0, end_mask = cache_77_end_mask_0, squeeze_mask = cache_77_squeeze_mask_0, x = value_3)[name = tensor("cache_77")]; + tensor cache_79_begin_0 = const()[name = tensor("cache_79_begin_0"), val = tensor([19, 0, 0, 0])]; + tensor cache_79_end_0 = const()[name = tensor("cache_79_end_0"), val = tensor([20, 1, 1024, 8])]; + tensor cache_79_end_mask_0 = const()[name = tensor("cache_79_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_79_squeeze_mask_0 = const()[name = tensor("cache_79_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_79 = slice_by_index(begin = cache_79_begin_0, end = cache_79_end_0, end_mask = cache_79_end_mask_0, squeeze_mask = cache_79_squeeze_mask_0, x = value_5)[name = tensor("cache_79")]; + tensor input_1015_axes_0 = const()[name = tensor("input_1015_axes_0"), val = tensor([-1])]; + tensor input_1015 = layer_norm(axes = input_1015_axes_0, beta = module_layers_19_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_19_norm_feed_forward1_weight, x = input_1013)[name = tensor("input_1015")]; + tensor input_1017 = linear(bias = linear_1_bias_0, weight = module_layers_19_feed_forward1_linear1_weight_quantized, x = input_1015)[name = tensor("linear_172")]; + tensor input_1019 = silu(x = input_1017)[name = tensor("input_1019")]; + tensor input_1023 = linear(bias = linear_2_bias_0, weight = module_layers_19_feed_forward1_linear2_weight_quantized, x = input_1019)[name = tensor("linear_173")]; + tensor var_4326 = const()[name = tensor("op_4326"), val = tensor(0x1p-1)]; + tensor var_4327 = mul(x = input_1023, y = var_4326)[name = tensor("op_4327")]; + tensor input_1025 = add(x = input_1013, y = var_4327)[name = tensor("input_1025")]; + tensor key_39_axes_0 = const()[name = tensor("key_39_axes_0"), val = tensor([-1])]; + tensor key_39 = layer_norm(axes = key_39_axes_0, beta = module_layers_19_norm_self_att_bias, epsilon = var_38, gamma = module_layers_19_norm_self_att_weight, x = input_1025)[name = tensor("key_39")]; + tensor input_1027_interleave_0 = const()[name = tensor("input_1027_interleave_0"), val = tensor(false)]; + tensor input_1027 = concat(axis = var_65, interleave = input_1027_interleave_0, values = (cache_77, key_39))[name = tensor("input_1027")]; + tensor var_4349_begin_0 = const()[name = tensor("op_4349_begin_0"), val = tensor([0, 2, 0])]; + tensor var_4349_end_0 = const()[name = tensor("op_4349_end_0"), val = tensor([1, 70, 1024])]; + tensor var_4349_end_mask_0 = const()[name = tensor("op_4349_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4349 = slice_by_index(begin = var_4349_begin_0, end = var_4349_end_0, end_mask = var_4349_end_mask_0, x = cache_77)[name = tensor("op_4349")]; + tensor var_4355_interleave_0 = const()[name = tensor("op_4355_interleave_0"), val = tensor(false)]; + tensor var_4355 = concat(axis = var_65, interleave = var_4355_interleave_0, values = (var_4349, key_39))[name = tensor("op_4355")]; + tensor var_4358 = linear(bias = linear_2_bias_0, weight = module_layers_19_self_attn_linear_q_weight_quantized, x = key_39)[name = tensor("linear_174")]; + tensor var_4359 = const()[name = tensor("op_4359"), val = tensor([1, -1, 8, 128])]; + tensor q_115 = reshape(shape = var_4359, x = var_4358)[name = tensor("q_115")]; + tensor var_4362 = linear(bias = linear_2_bias_0, weight = module_layers_19_self_attn_linear_k_weight_quantized, x = input_1027)[name = tensor("linear_175")]; + tensor var_4363 = const()[name = tensor("op_4363"), val = tensor([1, -1, 8, 128])]; + tensor k_77 = reshape(shape = var_4363, x = var_4362)[name = tensor("k_77")]; + tensor var_4366 = linear(bias = linear_2_bias_0, weight = module_layers_19_self_attn_linear_v_weight_quantized, x = input_1027)[name = tensor("linear_176")]; + tensor var_4367 = const()[name = tensor("op_4367"), val = tensor([1, -1, 8, 128])]; + tensor v_39 = reshape(shape = var_4367, x = var_4366)[name = tensor("v_39")]; + tensor value_47_perm_0 = const()[name = tensor("value_47_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_4379 = add(x = q_115, y = module_layers_19_self_attn_pos_bias_u)[name = tensor("op_4379")]; + tensor var_4381 = add(x = q_115, y = module_layers_19_self_attn_pos_bias_v)[name = tensor("op_4381")]; + tensor q_with_bias_v_39_perm_0 = const()[name = tensor("q_with_bias_v_39_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_4383_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4383_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590133632))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590280384))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590280128)))]; + tensor x_501_transpose_x_0 = const()[name = tensor("x_501_transpose_x_0"), val = tensor(false)]; + tensor x_501_transpose_y_0 = const()[name = tensor("x_501_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_39 = transpose(perm = q_with_bias_v_39_perm_0, x = var_4381)[name = tensor("transpose_191")]; + tensor x_501 = matmul(transpose_x = x_501_transpose_x_0, transpose_y = x_501_transpose_y_0, x = q_with_bias_v_39, y = op_4383_quantized)[name = tensor("x_501")]; + tensor const_326 = const()[name = tensor("const_326"), val = tensor(0x0p+0)]; + tensor x_503_pad_0 = const()[name = tensor("x_503_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_503_mode_0 = const()[name = tensor("x_503_mode_0"), val = tensor("constant")]; + tensor x_503 = pad(constant_val = const_326, mode = x_503_mode_0, pad = x_503_pad_0, x = x_501)[name = tensor("x_503")]; + tensor var_4391 = const()[name = tensor("op_4391"), val = tensor([1, 8, -1, 2])]; + tensor x_505 = reshape(shape = var_4391, x = x_503)[name = tensor("x_505")]; + tensor var_4395_begin_0 = const()[name = tensor("op_4395_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_4395_end_0 = const()[name = tensor("op_4395_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_4395_end_mask_0 = const()[name = tensor("op_4395_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_4395 = slice_by_index(begin = var_4395_begin_0, end = var_4395_end_0, end_mask = var_4395_end_mask_0, x = x_505)[name = tensor("op_4395")]; + tensor var_4396 = const()[name = tensor("op_4396"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_77 = reshape(shape = var_4396, x = var_4395)[name = tensor("matrix_bd_77")]; + tensor matrix_ac_39_transpose_x_0 = const()[name = tensor("matrix_ac_39_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_39_transpose_y_0 = const()[name = tensor("matrix_ac_39_transpose_y_0"), val = tensor(false)]; + tensor transpose_134_perm_0 = const()[name = tensor("transpose_134_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_135_perm_0 = const()[name = tensor("transpose_135_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_135 = transpose(perm = transpose_135_perm_0, x = k_77)[name = tensor("transpose_189")]; + tensor transpose_134 = transpose(perm = transpose_134_perm_0, x = var_4379)[name = tensor("transpose_190")]; + tensor matrix_ac_39 = matmul(transpose_x = matrix_ac_39_transpose_x_0, transpose_y = matrix_ac_39_transpose_y_0, x = transpose_134, y = transpose_135)[name = tensor("matrix_ac_39")]; + tensor matrix_bd_79_begin_0 = const()[name = tensor("matrix_bd_79_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_79_end_0 = const()[name = tensor("matrix_bd_79_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_79_end_mask_0 = const()[name = tensor("matrix_bd_79_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_79 = slice_by_index(begin = matrix_bd_79_begin_0, end = matrix_bd_79_end_0, end_mask = matrix_bd_79_end_mask_0, x = matrix_bd_77)[name = tensor("matrix_bd_79")]; + tensor var_4405 = add(x = matrix_ac_39, y = matrix_bd_79)[name = tensor("op_4405")]; + tensor _inversed_scores_77_y_0 = const()[name = tensor("_inversed_scores_77_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_77 = mul(x = var_4405, y = _inversed_scores_77_y_0)[name = tensor("_inversed_scores_77")]; + tensor scores_79 = select(a = var_41, b = _inversed_scores_77, cond = mask_11)[name = tensor("scores_79")]; + tensor var_4411 = softmax(axis = var_56, x = scores_79)[name = tensor("op_4411")]; + tensor input_1029 = select(a = var_40, b = var_4411, cond = mask_11)[name = tensor("input_1029")]; + tensor x_507_transpose_x_0 = const()[name = tensor("x_507_transpose_x_0"), val = tensor(false)]; + tensor x_507_transpose_y_0 = const()[name = tensor("x_507_transpose_y_0"), val = tensor(false)]; + tensor value_47 = transpose(perm = value_47_perm_0, x = v_39)[name = tensor("transpose_188")]; + tensor x_507 = matmul(transpose_x = x_507_transpose_x_0, transpose_y = x_507_transpose_y_0, x = input_1029, y = value_47)[name = tensor("x_507")]; + tensor var_4415_perm_0 = const()[name = tensor("op_4415_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_4416 = const()[name = tensor("op_4416"), val = tensor([1, -1, 1024])]; + tensor var_4415 = transpose(perm = var_4415_perm_0, x = x_507)[name = tensor("transpose_187")]; + tensor input_1031 = reshape(shape = var_4416, x = var_4415)[name = tensor("input_1031")]; + tensor input_1033 = linear(bias = linear_2_bias_0, weight = module_layers_19_self_attn_linear_out_weight_quantized, x = input_1031)[name = tensor("linear_178")]; + tensor input_1035 = add(x = input_1025, y = input_1033)[name = tensor("input_1035")]; + tensor x_511_axes_0 = const()[name = tensor("x_511_axes_0"), val = tensor([-1])]; + tensor x_511 = layer_norm(axes = x_511_axes_0, beta = module_layers_19_norm_conv_bias, epsilon = var_38, gamma = module_layers_19_norm_conv_weight, x = input_1035)[name = tensor("x_511")]; + tensor input_1037_perm_0 = const()[name = tensor("input_1037_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1039_pad_type_0 = const()[name = tensor("input_1039_pad_type_0"), val = tensor("valid")]; + tensor input_1039_strides_0 = const()[name = tensor("input_1039_strides_0"), val = tensor([1])]; + tensor input_1039_pad_0 = const()[name = tensor("input_1039_pad_0"), val = tensor([0, 0])]; + tensor input_1039_dilations_0 = const()[name = tensor("input_1039_dilations_0"), val = tensor([1])]; + tensor input_1039_groups_0 = const()[name = tensor("input_1039_groups_0"), val = tensor(1)]; + tensor input_1037 = transpose(perm = input_1037_perm_0, x = x_511)[name = tensor("transpose_186")]; + tensor input_1039 = conv(dilations = input_1039_dilations_0, groups = input_1039_groups_0, pad = input_1039_pad_0, pad_type = input_1039_pad_type_0, strides = input_1039_strides_0, weight = module_layers_19_conv_pointwise_conv1_weight_quantized, x = input_1037)[name = tensor("input_1039")]; + tensor x_513_split_num_splits_0 = const()[name = tensor("x_513_split_num_splits_0"), val = tensor(2)]; + tensor x_513_split_axis_0 = const()[name = tensor("x_513_split_axis_0"), val = tensor(1)]; + tensor x_513_split_0, tensor x_513_split_1 = split(axis = x_513_split_axis_0, num_splits = x_513_split_num_splits_0, x = input_1039)[name = tensor("x_513_split")]; + tensor x_513_split_1_sigmoid = sigmoid(x = x_513_split_1)[name = tensor("x_513_split_1_sigmoid")]; + tensor x_513 = mul(x = x_513_split_0, y = x_513_split_1_sigmoid)[name = tensor("x_513")]; + tensor input_1041 = select(a = var_40, b = x_513, cond = var_565)[name = tensor("input_1041")]; + tensor new_x_79_interleave_0 = const()[name = tensor("new_x_79_interleave_0"), val = tensor(false)]; + tensor new_x_79 = concat(axis = var_56, interleave = new_x_79_interleave_0, values = (cache_79, input_1041))[name = tensor("new_x_79")]; + tensor var_4454_begin_0 = const()[name = tensor("op_4454_begin_0"), val = tensor([0, 0, 2])]; + tensor var_4454_end_0 = const()[name = tensor("op_4454_end_0"), val = tensor([1, 1024, 10])]; + tensor var_4454_end_mask_0 = const()[name = tensor("op_4454_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4454 = slice_by_index(begin = var_4454_begin_0, end = var_4454_end_0, end_mask = var_4454_end_mask_0, x = new_x_79)[name = tensor("op_4454")]; + tensor x_515_pad_type_0 = const()[name = tensor("x_515_pad_type_0"), val = tensor("valid")]; + tensor x_515_groups_0 = const()[name = tensor("x_515_groups_0"), val = tensor(1024)]; + tensor x_515_strides_0 = const()[name = tensor("x_515_strides_0"), val = tensor([1])]; + tensor x_515_pad_0 = const()[name = tensor("x_515_pad_0"), val = tensor([0, 0])]; + tensor x_515_dilations_0 = const()[name = tensor("x_515_dilations_0"), val = tensor([1])]; + tensor x_515 = conv(dilations = x_515_dilations_0, groups = x_515_groups_0, pad = x_515_pad_0, pad_type = x_515_pad_type_0, strides = x_515_strides_0, weight = module_layers_19_conv_depthwise_conv_weight_quantized, x = new_x_79)[name = tensor("x_515")]; + tensor input_1043_perm_0 = const()[name = tensor("input_1043_perm_0"), val = tensor([0, 2, 1])]; + tensor x_517_axes_0 = const()[name = tensor("x_517_axes_0"), val = tensor([-1])]; + tensor input_1043 = transpose(perm = input_1043_perm_0, x = x_515)[name = tensor("transpose_185")]; + tensor x_517 = layer_norm(axes = x_517_axes_0, beta = module_layers_19_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_19_conv_batch_norm_weight, x = input_1043)[name = tensor("x_517")]; + tensor input_1045_perm_0 = const()[name = tensor("input_1045_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1045 = transpose(perm = input_1045_perm_0, x = x_517)[name = tensor("transpose_184")]; + tensor input_1047 = silu(x = input_1045)[name = tensor("input_1047")]; + tensor x_519_pad_type_0 = const()[name = tensor("x_519_pad_type_0"), val = tensor("valid")]; + tensor x_519_strides_0 = const()[name = tensor("x_519_strides_0"), val = tensor([1])]; + tensor x_519_pad_0 = const()[name = tensor("x_519_pad_0"), val = tensor([0, 0])]; + tensor x_519_dilations_0 = const()[name = tensor("x_519_dilations_0"), val = tensor([1])]; + tensor x_519_groups_0 = const()[name = tensor("x_519_groups_0"), val = tensor(1)]; + tensor x_519 = conv(dilations = x_519_dilations_0, groups = x_519_groups_0, pad = x_519_pad_0, pad_type = x_519_pad_type_0, strides = x_519_strides_0, weight = module_layers_19_conv_pointwise_conv2_weight_quantized, x = input_1047)[name = tensor("x_519")]; + tensor input_1049_perm_0 = const()[name = tensor("input_1049_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1049 = transpose(perm = input_1049_perm_0, x = x_519)[name = tensor("transpose_183")]; + tensor input_1051 = add(x = input_1035, y = input_1049)[name = tensor("input_1051")]; + tensor input_1053_axes_0 = const()[name = tensor("input_1053_axes_0"), val = tensor([-1])]; + tensor input_1053 = layer_norm(axes = input_1053_axes_0, beta = module_layers_19_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_19_norm_feed_forward2_weight, x = input_1051)[name = tensor("input_1053")]; + tensor input_1055 = linear(bias = linear_1_bias_0, weight = module_layers_19_feed_forward2_linear1_weight_quantized, x = input_1053)[name = tensor("linear_179")]; + tensor input_1057 = silu(x = input_1055)[name = tensor("input_1057")]; + tensor input_1061 = linear(bias = linear_2_bias_0, weight = module_layers_19_feed_forward2_linear2_weight_quantized, x = input_1057)[name = tensor("linear_180")]; + tensor var_4495 = const()[name = tensor("op_4495"), val = tensor(0x1p-1)]; + tensor var_4496 = mul(x = input_1061, y = var_4495)[name = tensor("op_4496")]; + tensor input_1063 = add(x = input_1051, y = var_4496)[name = tensor("input_1063")]; + tensor input_1065_axes_0 = const()[name = tensor("input_1065_axes_0"), val = tensor([-1])]; + tensor input_1065 = layer_norm(axes = input_1065_axes_0, beta = module_layers_19_norm_out_bias, epsilon = var_38, gamma = module_layers_19_norm_out_weight, x = input_1063)[name = tensor("input_1065")]; + tensor cache_81_begin_0 = const()[name = tensor("cache_81_begin_0"), val = tensor([20, 0, 0, 0])]; + tensor cache_81_end_0 = const()[name = tensor("cache_81_end_0"), val = tensor([21, 1, 70, 1024])]; + tensor cache_81_end_mask_0 = const()[name = tensor("cache_81_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_81_squeeze_mask_0 = const()[name = tensor("cache_81_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_81 = slice_by_index(begin = cache_81_begin_0, end = cache_81_end_0, end_mask = cache_81_end_mask_0, squeeze_mask = cache_81_squeeze_mask_0, x = value_3)[name = tensor("cache_81")]; + tensor cache_83_begin_0 = const()[name = tensor("cache_83_begin_0"), val = tensor([20, 0, 0, 0])]; + tensor cache_83_end_0 = const()[name = tensor("cache_83_end_0"), val = tensor([21, 1, 1024, 8])]; + tensor cache_83_end_mask_0 = const()[name = tensor("cache_83_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_83_squeeze_mask_0 = const()[name = tensor("cache_83_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_83 = slice_by_index(begin = cache_83_begin_0, end = cache_83_end_0, end_mask = cache_83_end_mask_0, squeeze_mask = cache_83_squeeze_mask_0, x = value_5)[name = tensor("cache_83")]; + tensor input_1067_axes_0 = const()[name = tensor("input_1067_axes_0"), val = tensor([-1])]; + tensor input_1067 = layer_norm(axes = input_1067_axes_0, beta = module_layers_20_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_20_norm_feed_forward1_weight, x = input_1065)[name = tensor("input_1067")]; + tensor input_1069 = linear(bias = linear_1_bias_0, weight = module_layers_20_feed_forward1_linear1_weight_quantized, x = input_1067)[name = tensor("linear_181")]; + tensor input_1071 = silu(x = input_1069)[name = tensor("input_1071")]; + tensor input_1075 = linear(bias = linear_2_bias_0, weight = module_layers_20_feed_forward1_linear2_weight_quantized, x = input_1071)[name = tensor("linear_182")]; + tensor var_4530 = const()[name = tensor("op_4530"), val = tensor(0x1p-1)]; + tensor var_4531 = mul(x = input_1075, y = var_4530)[name = tensor("op_4531")]; + tensor input_1077 = add(x = input_1065, y = var_4531)[name = tensor("input_1077")]; + tensor key_41_axes_0 = const()[name = tensor("key_41_axes_0"), val = tensor([-1])]; + tensor key_41 = layer_norm(axes = key_41_axes_0, beta = module_layers_20_norm_self_att_bias, epsilon = var_38, gamma = module_layers_20_norm_self_att_weight, x = input_1077)[name = tensor("key_41")]; + tensor input_1079_interleave_0 = const()[name = tensor("input_1079_interleave_0"), val = tensor(false)]; + tensor input_1079 = concat(axis = var_65, interleave = input_1079_interleave_0, values = (cache_81, key_41))[name = tensor("input_1079")]; + tensor var_4553_begin_0 = const()[name = tensor("op_4553_begin_0"), val = tensor([0, 2, 0])]; + tensor var_4553_end_0 = const()[name = tensor("op_4553_end_0"), val = tensor([1, 70, 1024])]; + tensor var_4553_end_mask_0 = const()[name = tensor("op_4553_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4553 = slice_by_index(begin = var_4553_begin_0, end = var_4553_end_0, end_mask = var_4553_end_mask_0, x = cache_81)[name = tensor("op_4553")]; + tensor var_4559_interleave_0 = const()[name = tensor("op_4559_interleave_0"), val = tensor(false)]; + tensor var_4559 = concat(axis = var_65, interleave = var_4559_interleave_0, values = (var_4553, key_41))[name = tensor("op_4559")]; + tensor var_4562 = linear(bias = linear_2_bias_0, weight = module_layers_20_self_attn_linear_q_weight_quantized, x = key_41)[name = tensor("linear_183")]; + tensor var_4563 = const()[name = tensor("op_4563"), val = tensor([1, -1, 8, 128])]; + tensor q_121 = reshape(shape = var_4563, x = var_4562)[name = tensor("q_121")]; + tensor var_4566 = linear(bias = linear_2_bias_0, weight = module_layers_20_self_attn_linear_k_weight_quantized, x = input_1079)[name = tensor("linear_184")]; + tensor var_4567 = const()[name = tensor("op_4567"), val = tensor([1, -1, 8, 128])]; + tensor k_81 = reshape(shape = var_4567, x = var_4566)[name = tensor("k_81")]; + tensor var_4570 = linear(bias = linear_2_bias_0, weight = module_layers_20_self_attn_linear_v_weight_quantized, x = input_1079)[name = tensor("linear_185")]; + tensor var_4571 = const()[name = tensor("op_4571"), val = tensor([1, -1, 8, 128])]; + tensor v_41 = reshape(shape = var_4571, x = var_4570)[name = tensor("v_41")]; + tensor value_49_perm_0 = const()[name = tensor("value_49_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_4583 = add(x = q_121, y = module_layers_20_self_attn_pos_bias_u)[name = tensor("op_4583")]; + tensor var_4585 = add(x = q_121, y = module_layers_20_self_attn_pos_bias_v)[name = tensor("op_4585")]; + tensor q_with_bias_v_41_perm_0 = const()[name = tensor("q_with_bias_v_41_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_4587_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4587_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590281024))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590427776))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590427520)))]; + tensor x_527_transpose_x_0 = const()[name = tensor("x_527_transpose_x_0"), val = tensor(false)]; + tensor x_527_transpose_y_0 = const()[name = tensor("x_527_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_41 = transpose(perm = q_with_bias_v_41_perm_0, x = var_4585)[name = tensor("transpose_182")]; + tensor x_527 = matmul(transpose_x = x_527_transpose_x_0, transpose_y = x_527_transpose_y_0, x = q_with_bias_v_41, y = op_4587_quantized)[name = tensor("x_527")]; + tensor const_339 = const()[name = tensor("const_339"), val = tensor(0x0p+0)]; + tensor x_529_pad_0 = const()[name = tensor("x_529_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_529_mode_0 = const()[name = tensor("x_529_mode_0"), val = tensor("constant")]; + tensor x_529 = pad(constant_val = const_339, mode = x_529_mode_0, pad = x_529_pad_0, x = x_527)[name = tensor("x_529")]; + tensor var_4595 = const()[name = tensor("op_4595"), val = tensor([1, 8, -1, 2])]; + tensor x_531 = reshape(shape = var_4595, x = x_529)[name = tensor("x_531")]; + tensor var_4599_begin_0 = const()[name = tensor("op_4599_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_4599_end_0 = const()[name = tensor("op_4599_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_4599_end_mask_0 = const()[name = tensor("op_4599_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_4599 = slice_by_index(begin = var_4599_begin_0, end = var_4599_end_0, end_mask = var_4599_end_mask_0, x = x_531)[name = tensor("op_4599")]; + tensor var_4600 = const()[name = tensor("op_4600"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_81 = reshape(shape = var_4600, x = var_4599)[name = tensor("matrix_bd_81")]; + tensor matrix_ac_41_transpose_x_0 = const()[name = tensor("matrix_ac_41_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_41_transpose_y_0 = const()[name = tensor("matrix_ac_41_transpose_y_0"), val = tensor(false)]; + tensor transpose_136_perm_0 = const()[name = tensor("transpose_136_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_137_perm_0 = const()[name = tensor("transpose_137_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_137 = transpose(perm = transpose_137_perm_0, x = k_81)[name = tensor("transpose_180")]; + tensor transpose_136 = transpose(perm = transpose_136_perm_0, x = var_4583)[name = tensor("transpose_181")]; + tensor matrix_ac_41 = matmul(transpose_x = matrix_ac_41_transpose_x_0, transpose_y = matrix_ac_41_transpose_y_0, x = transpose_136, y = transpose_137)[name = tensor("matrix_ac_41")]; + tensor matrix_bd_83_begin_0 = const()[name = tensor("matrix_bd_83_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_83_end_0 = const()[name = tensor("matrix_bd_83_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_83_end_mask_0 = const()[name = tensor("matrix_bd_83_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_83 = slice_by_index(begin = matrix_bd_83_begin_0, end = matrix_bd_83_end_0, end_mask = matrix_bd_83_end_mask_0, x = matrix_bd_81)[name = tensor("matrix_bd_83")]; + tensor var_4609 = add(x = matrix_ac_41, y = matrix_bd_83)[name = tensor("op_4609")]; + tensor _inversed_scores_81_y_0 = const()[name = tensor("_inversed_scores_81_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_81 = mul(x = var_4609, y = _inversed_scores_81_y_0)[name = tensor("_inversed_scores_81")]; + tensor scores_83 = select(a = var_41, b = _inversed_scores_81, cond = mask_11)[name = tensor("scores_83")]; + tensor var_4615 = softmax(axis = var_56, x = scores_83)[name = tensor("op_4615")]; + tensor input_1081 = select(a = var_40, b = var_4615, cond = mask_11)[name = tensor("input_1081")]; + tensor x_533_transpose_x_0 = const()[name = tensor("x_533_transpose_x_0"), val = tensor(false)]; + tensor x_533_transpose_y_0 = const()[name = tensor("x_533_transpose_y_0"), val = tensor(false)]; + tensor value_49 = transpose(perm = value_49_perm_0, x = v_41)[name = tensor("transpose_179")]; + tensor x_533 = matmul(transpose_x = x_533_transpose_x_0, transpose_y = x_533_transpose_y_0, x = input_1081, y = value_49)[name = tensor("x_533")]; + tensor var_4619_perm_0 = const()[name = tensor("op_4619_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_4620 = const()[name = tensor("op_4620"), val = tensor([1, -1, 1024])]; + tensor var_4619 = transpose(perm = var_4619_perm_0, x = x_533)[name = tensor("transpose_178")]; + tensor input_1083 = reshape(shape = var_4620, x = var_4619)[name = tensor("input_1083")]; + tensor input_1085 = linear(bias = linear_2_bias_0, weight = module_layers_20_self_attn_linear_out_weight_quantized, x = input_1083)[name = tensor("linear_187")]; + tensor input_1087 = add(x = input_1077, y = input_1085)[name = tensor("input_1087")]; + tensor x_537_axes_0 = const()[name = tensor("x_537_axes_0"), val = tensor([-1])]; + tensor x_537 = layer_norm(axes = x_537_axes_0, beta = module_layers_20_norm_conv_bias, epsilon = var_38, gamma = module_layers_20_norm_conv_weight, x = input_1087)[name = tensor("x_537")]; + tensor input_1089_perm_0 = const()[name = tensor("input_1089_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1091_pad_type_0 = const()[name = tensor("input_1091_pad_type_0"), val = tensor("valid")]; + tensor input_1091_strides_0 = const()[name = tensor("input_1091_strides_0"), val = tensor([1])]; + tensor input_1091_pad_0 = const()[name = tensor("input_1091_pad_0"), val = tensor([0, 0])]; + tensor input_1091_dilations_0 = const()[name = tensor("input_1091_dilations_0"), val = tensor([1])]; + tensor input_1091_groups_0 = const()[name = tensor("input_1091_groups_0"), val = tensor(1)]; + tensor input_1089 = transpose(perm = input_1089_perm_0, x = x_537)[name = tensor("transpose_177")]; + tensor input_1091 = conv(dilations = input_1091_dilations_0, groups = input_1091_groups_0, pad = input_1091_pad_0, pad_type = input_1091_pad_type_0, strides = input_1091_strides_0, weight = module_layers_20_conv_pointwise_conv1_weight_quantized, x = input_1089)[name = tensor("input_1091")]; + tensor x_539_split_num_splits_0 = const()[name = tensor("x_539_split_num_splits_0"), val = tensor(2)]; + tensor x_539_split_axis_0 = const()[name = tensor("x_539_split_axis_0"), val = tensor(1)]; + tensor x_539_split_0, tensor x_539_split_1 = split(axis = x_539_split_axis_0, num_splits = x_539_split_num_splits_0, x = input_1091)[name = tensor("x_539_split")]; + tensor x_539_split_1_sigmoid = sigmoid(x = x_539_split_1)[name = tensor("x_539_split_1_sigmoid")]; + tensor x_539 = mul(x = x_539_split_0, y = x_539_split_1_sigmoid)[name = tensor("x_539")]; + tensor input_1093 = select(a = var_40, b = x_539, cond = var_565)[name = tensor("input_1093")]; + tensor new_x_83_interleave_0 = const()[name = tensor("new_x_83_interleave_0"), val = tensor(false)]; + tensor new_x_83 = concat(axis = var_56, interleave = new_x_83_interleave_0, values = (cache_83, input_1093))[name = tensor("new_x_83")]; + tensor var_4658_begin_0 = const()[name = tensor("op_4658_begin_0"), val = tensor([0, 0, 2])]; + tensor var_4658_end_0 = const()[name = tensor("op_4658_end_0"), val = tensor([1, 1024, 10])]; + tensor var_4658_end_mask_0 = const()[name = tensor("op_4658_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4658 = slice_by_index(begin = var_4658_begin_0, end = var_4658_end_0, end_mask = var_4658_end_mask_0, x = new_x_83)[name = tensor("op_4658")]; + tensor x_541_pad_type_0 = const()[name = tensor("x_541_pad_type_0"), val = tensor("valid")]; + tensor x_541_groups_0 = const()[name = tensor("x_541_groups_0"), val = tensor(1024)]; + tensor x_541_strides_0 = const()[name = tensor("x_541_strides_0"), val = tensor([1])]; + tensor x_541_pad_0 = const()[name = tensor("x_541_pad_0"), val = tensor([0, 0])]; + tensor x_541_dilations_0 = const()[name = tensor("x_541_dilations_0"), val = tensor([1])]; + tensor x_541 = conv(dilations = x_541_dilations_0, groups = x_541_groups_0, pad = x_541_pad_0, pad_type = x_541_pad_type_0, strides = x_541_strides_0, weight = module_layers_20_conv_depthwise_conv_weight_quantized, x = new_x_83)[name = tensor("x_541")]; + tensor input_1095_perm_0 = const()[name = tensor("input_1095_perm_0"), val = tensor([0, 2, 1])]; + tensor x_543_axes_0 = const()[name = tensor("x_543_axes_0"), val = tensor([-1])]; + tensor input_1095 = transpose(perm = input_1095_perm_0, x = x_541)[name = tensor("transpose_176")]; + tensor x_543 = layer_norm(axes = x_543_axes_0, beta = module_layers_20_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_20_conv_batch_norm_weight, x = input_1095)[name = tensor("x_543")]; + tensor input_1097_perm_0 = const()[name = tensor("input_1097_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1097 = transpose(perm = input_1097_perm_0, x = x_543)[name = tensor("transpose_175")]; + tensor input_1099 = silu(x = input_1097)[name = tensor("input_1099")]; + tensor x_545_pad_type_0 = const()[name = tensor("x_545_pad_type_0"), val = tensor("valid")]; + tensor x_545_strides_0 = const()[name = tensor("x_545_strides_0"), val = tensor([1])]; + tensor x_545_pad_0 = const()[name = tensor("x_545_pad_0"), val = tensor([0, 0])]; + tensor x_545_dilations_0 = const()[name = tensor("x_545_dilations_0"), val = tensor([1])]; + tensor x_545_groups_0 = const()[name = tensor("x_545_groups_0"), val = tensor(1)]; + tensor x_545 = conv(dilations = x_545_dilations_0, groups = x_545_groups_0, pad = x_545_pad_0, pad_type = x_545_pad_type_0, strides = x_545_strides_0, weight = module_layers_20_conv_pointwise_conv2_weight_quantized, x = input_1099)[name = tensor("x_545")]; + tensor input_1101_perm_0 = const()[name = tensor("input_1101_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1101 = transpose(perm = input_1101_perm_0, x = x_545)[name = tensor("transpose_174")]; + tensor input_1103 = add(x = input_1087, y = input_1101)[name = tensor("input_1103")]; + tensor input_1105_axes_0 = const()[name = tensor("input_1105_axes_0"), val = tensor([-1])]; + tensor input_1105 = layer_norm(axes = input_1105_axes_0, beta = module_layers_20_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_20_norm_feed_forward2_weight, x = input_1103)[name = tensor("input_1105")]; + tensor input_1107 = linear(bias = linear_1_bias_0, weight = module_layers_20_feed_forward2_linear1_weight_quantized, x = input_1105)[name = tensor("linear_188")]; + tensor input_1109 = silu(x = input_1107)[name = tensor("input_1109")]; + tensor input_1113 = linear(bias = linear_2_bias_0, weight = module_layers_20_feed_forward2_linear2_weight_quantized, x = input_1109)[name = tensor("linear_189")]; + tensor var_4699 = const()[name = tensor("op_4699"), val = tensor(0x1p-1)]; + tensor var_4700 = mul(x = input_1113, y = var_4699)[name = tensor("op_4700")]; + tensor input_1115 = add(x = input_1103, y = var_4700)[name = tensor("input_1115")]; + tensor input_1117_axes_0 = const()[name = tensor("input_1117_axes_0"), val = tensor([-1])]; + tensor input_1117 = layer_norm(axes = input_1117_axes_0, beta = module_layers_20_norm_out_bias, epsilon = var_38, gamma = module_layers_20_norm_out_weight, x = input_1115)[name = tensor("input_1117")]; + tensor cache_85_begin_0 = const()[name = tensor("cache_85_begin_0"), val = tensor([21, 0, 0, 0])]; + tensor cache_85_end_0 = const()[name = tensor("cache_85_end_0"), val = tensor([22, 1, 70, 1024])]; + tensor cache_85_end_mask_0 = const()[name = tensor("cache_85_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_85_squeeze_mask_0 = const()[name = tensor("cache_85_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_85 = slice_by_index(begin = cache_85_begin_0, end = cache_85_end_0, end_mask = cache_85_end_mask_0, squeeze_mask = cache_85_squeeze_mask_0, x = value_3)[name = tensor("cache_85")]; + tensor cache_87_begin_0 = const()[name = tensor("cache_87_begin_0"), val = tensor([21, 0, 0, 0])]; + tensor cache_87_end_0 = const()[name = tensor("cache_87_end_0"), val = tensor([22, 1, 1024, 8])]; + tensor cache_87_end_mask_0 = const()[name = tensor("cache_87_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_87_squeeze_mask_0 = const()[name = tensor("cache_87_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_87 = slice_by_index(begin = cache_87_begin_0, end = cache_87_end_0, end_mask = cache_87_end_mask_0, squeeze_mask = cache_87_squeeze_mask_0, x = value_5)[name = tensor("cache_87")]; + tensor input_1119_axes_0 = const()[name = tensor("input_1119_axes_0"), val = tensor([-1])]; + tensor input_1119 = layer_norm(axes = input_1119_axes_0, beta = module_layers_21_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_21_norm_feed_forward1_weight, x = input_1117)[name = tensor("input_1119")]; + tensor input_1121 = linear(bias = linear_1_bias_0, weight = module_layers_21_feed_forward1_linear1_weight_quantized, x = input_1119)[name = tensor("linear_190")]; + tensor input_1123 = silu(x = input_1121)[name = tensor("input_1123")]; + tensor input_1127 = linear(bias = linear_2_bias_0, weight = module_layers_21_feed_forward1_linear2_weight_quantized, x = input_1123)[name = tensor("linear_191")]; + tensor var_4734 = const()[name = tensor("op_4734"), val = tensor(0x1p-1)]; + tensor var_4735 = mul(x = input_1127, y = var_4734)[name = tensor("op_4735")]; + tensor input_1129 = add(x = input_1117, y = var_4735)[name = tensor("input_1129")]; + tensor key_43_axes_0 = const()[name = tensor("key_43_axes_0"), val = tensor([-1])]; + tensor key_43 = layer_norm(axes = key_43_axes_0, beta = module_layers_21_norm_self_att_bias, epsilon = var_38, gamma = module_layers_21_norm_self_att_weight, x = input_1129)[name = tensor("key_43")]; + tensor input_1131_interleave_0 = const()[name = tensor("input_1131_interleave_0"), val = tensor(false)]; + tensor input_1131 = concat(axis = var_65, interleave = input_1131_interleave_0, values = (cache_85, key_43))[name = tensor("input_1131")]; + tensor var_4757_begin_0 = const()[name = tensor("op_4757_begin_0"), val = tensor([0, 2, 0])]; + tensor var_4757_end_0 = const()[name = tensor("op_4757_end_0"), val = tensor([1, 70, 1024])]; + tensor var_4757_end_mask_0 = const()[name = tensor("op_4757_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4757 = slice_by_index(begin = var_4757_begin_0, end = var_4757_end_0, end_mask = var_4757_end_mask_0, x = cache_85)[name = tensor("op_4757")]; + tensor var_4763_interleave_0 = const()[name = tensor("op_4763_interleave_0"), val = tensor(false)]; + tensor var_4763 = concat(axis = var_65, interleave = var_4763_interleave_0, values = (var_4757, key_43))[name = tensor("op_4763")]; + tensor var_4766 = linear(bias = linear_2_bias_0, weight = module_layers_21_self_attn_linear_q_weight_quantized, x = key_43)[name = tensor("linear_192")]; + tensor var_4767 = const()[name = tensor("op_4767"), val = tensor([1, -1, 8, 128])]; + tensor q_127 = reshape(shape = var_4767, x = var_4766)[name = tensor("q_127")]; + tensor var_4770 = linear(bias = linear_2_bias_0, weight = module_layers_21_self_attn_linear_k_weight_quantized, x = input_1131)[name = tensor("linear_193")]; + tensor var_4771 = const()[name = tensor("op_4771"), val = tensor([1, -1, 8, 128])]; + tensor k_85 = reshape(shape = var_4771, x = var_4770)[name = tensor("k_85")]; + tensor var_4774 = linear(bias = linear_2_bias_0, weight = module_layers_21_self_attn_linear_v_weight_quantized, x = input_1131)[name = tensor("linear_194")]; + tensor var_4775 = const()[name = tensor("op_4775"), val = tensor([1, -1, 8, 128])]; + tensor v_43 = reshape(shape = var_4775, x = var_4774)[name = tensor("v_43")]; + tensor value_51_perm_0 = const()[name = tensor("value_51_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_4787 = add(x = q_127, y = module_layers_21_self_attn_pos_bias_u)[name = tensor("op_4787")]; + tensor var_4789 = add(x = q_127, y = module_layers_21_self_attn_pos_bias_v)[name = tensor("op_4789")]; + tensor q_with_bias_v_43_perm_0 = const()[name = tensor("q_with_bias_v_43_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_4791_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4791_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590428416))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590575168))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590574912)))]; + tensor x_553_transpose_x_0 = const()[name = tensor("x_553_transpose_x_0"), val = tensor(false)]; + tensor x_553_transpose_y_0 = const()[name = tensor("x_553_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_43 = transpose(perm = q_with_bias_v_43_perm_0, x = var_4789)[name = tensor("transpose_173")]; + tensor x_553 = matmul(transpose_x = x_553_transpose_x_0, transpose_y = x_553_transpose_y_0, x = q_with_bias_v_43, y = op_4791_quantized)[name = tensor("x_553")]; + tensor const_352 = const()[name = tensor("const_352"), val = tensor(0x0p+0)]; + tensor x_555_pad_0 = const()[name = tensor("x_555_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_555_mode_0 = const()[name = tensor("x_555_mode_0"), val = tensor("constant")]; + tensor x_555 = pad(constant_val = const_352, mode = x_555_mode_0, pad = x_555_pad_0, x = x_553)[name = tensor("x_555")]; + tensor var_4799 = const()[name = tensor("op_4799"), val = tensor([1, 8, -1, 2])]; + tensor x_557 = reshape(shape = var_4799, x = x_555)[name = tensor("x_557")]; + tensor var_4803_begin_0 = const()[name = tensor("op_4803_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_4803_end_0 = const()[name = tensor("op_4803_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_4803_end_mask_0 = const()[name = tensor("op_4803_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_4803 = slice_by_index(begin = var_4803_begin_0, end = var_4803_end_0, end_mask = var_4803_end_mask_0, x = x_557)[name = tensor("op_4803")]; + tensor var_4804 = const()[name = tensor("op_4804"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_85 = reshape(shape = var_4804, x = var_4803)[name = tensor("matrix_bd_85")]; + tensor matrix_ac_43_transpose_x_0 = const()[name = tensor("matrix_ac_43_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_43_transpose_y_0 = const()[name = tensor("matrix_ac_43_transpose_y_0"), val = tensor(false)]; + tensor transpose_138_perm_0 = const()[name = tensor("transpose_138_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_139_perm_0 = const()[name = tensor("transpose_139_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_139 = transpose(perm = transpose_139_perm_0, x = k_85)[name = tensor("transpose_171")]; + tensor transpose_138 = transpose(perm = transpose_138_perm_0, x = var_4787)[name = tensor("transpose_172")]; + tensor matrix_ac_43 = matmul(transpose_x = matrix_ac_43_transpose_x_0, transpose_y = matrix_ac_43_transpose_y_0, x = transpose_138, y = transpose_139)[name = tensor("matrix_ac_43")]; + tensor matrix_bd_87_begin_0 = const()[name = tensor("matrix_bd_87_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_87_end_0 = const()[name = tensor("matrix_bd_87_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_87_end_mask_0 = const()[name = tensor("matrix_bd_87_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_87 = slice_by_index(begin = matrix_bd_87_begin_0, end = matrix_bd_87_end_0, end_mask = matrix_bd_87_end_mask_0, x = matrix_bd_85)[name = tensor("matrix_bd_87")]; + tensor var_4813 = add(x = matrix_ac_43, y = matrix_bd_87)[name = tensor("op_4813")]; + tensor _inversed_scores_85_y_0 = const()[name = tensor("_inversed_scores_85_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_85 = mul(x = var_4813, y = _inversed_scores_85_y_0)[name = tensor("_inversed_scores_85")]; + tensor scores_87 = select(a = var_41, b = _inversed_scores_85, cond = mask_11)[name = tensor("scores_87")]; + tensor var_4819 = softmax(axis = var_56, x = scores_87)[name = tensor("op_4819")]; + tensor input_1133 = select(a = var_40, b = var_4819, cond = mask_11)[name = tensor("input_1133")]; + tensor x_559_transpose_x_0 = const()[name = tensor("x_559_transpose_x_0"), val = tensor(false)]; + tensor x_559_transpose_y_0 = const()[name = tensor("x_559_transpose_y_0"), val = tensor(false)]; + tensor value_51 = transpose(perm = value_51_perm_0, x = v_43)[name = tensor("transpose_170")]; + tensor x_559 = matmul(transpose_x = x_559_transpose_x_0, transpose_y = x_559_transpose_y_0, x = input_1133, y = value_51)[name = tensor("x_559")]; + tensor var_4823_perm_0 = const()[name = tensor("op_4823_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_4824 = const()[name = tensor("op_4824"), val = tensor([1, -1, 1024])]; + tensor var_4823 = transpose(perm = var_4823_perm_0, x = x_559)[name = tensor("transpose_169")]; + tensor input_1135 = reshape(shape = var_4824, x = var_4823)[name = tensor("input_1135")]; + tensor input_1137 = linear(bias = linear_2_bias_0, weight = module_layers_21_self_attn_linear_out_weight_quantized, x = input_1135)[name = tensor("linear_196")]; + tensor input_1139 = add(x = input_1129, y = input_1137)[name = tensor("input_1139")]; + tensor x_563_axes_0 = const()[name = tensor("x_563_axes_0"), val = tensor([-1])]; + tensor x_563 = layer_norm(axes = x_563_axes_0, beta = module_layers_21_norm_conv_bias, epsilon = var_38, gamma = module_layers_21_norm_conv_weight, x = input_1139)[name = tensor("x_563")]; + tensor input_1141_perm_0 = const()[name = tensor("input_1141_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1143_pad_type_0 = const()[name = tensor("input_1143_pad_type_0"), val = tensor("valid")]; + tensor input_1143_strides_0 = const()[name = tensor("input_1143_strides_0"), val = tensor([1])]; + tensor input_1143_pad_0 = const()[name = tensor("input_1143_pad_0"), val = tensor([0, 0])]; + tensor input_1143_dilations_0 = const()[name = tensor("input_1143_dilations_0"), val = tensor([1])]; + tensor input_1143_groups_0 = const()[name = tensor("input_1143_groups_0"), val = tensor(1)]; + tensor input_1141 = transpose(perm = input_1141_perm_0, x = x_563)[name = tensor("transpose_168")]; + tensor input_1143 = conv(dilations = input_1143_dilations_0, groups = input_1143_groups_0, pad = input_1143_pad_0, pad_type = input_1143_pad_type_0, strides = input_1143_strides_0, weight = module_layers_21_conv_pointwise_conv1_weight_quantized, x = input_1141)[name = tensor("input_1143")]; + tensor x_565_split_num_splits_0 = const()[name = tensor("x_565_split_num_splits_0"), val = tensor(2)]; + tensor x_565_split_axis_0 = const()[name = tensor("x_565_split_axis_0"), val = tensor(1)]; + tensor x_565_split_0, tensor x_565_split_1 = split(axis = x_565_split_axis_0, num_splits = x_565_split_num_splits_0, x = input_1143)[name = tensor("x_565_split")]; + tensor x_565_split_1_sigmoid = sigmoid(x = x_565_split_1)[name = tensor("x_565_split_1_sigmoid")]; + tensor x_565 = mul(x = x_565_split_0, y = x_565_split_1_sigmoid)[name = tensor("x_565")]; + tensor input_1145 = select(a = var_40, b = x_565, cond = var_565)[name = tensor("input_1145")]; + tensor new_x_87_interleave_0 = const()[name = tensor("new_x_87_interleave_0"), val = tensor(false)]; + tensor new_x_87 = concat(axis = var_56, interleave = new_x_87_interleave_0, values = (cache_87, input_1145))[name = tensor("new_x_87")]; + tensor var_4862_begin_0 = const()[name = tensor("op_4862_begin_0"), val = tensor([0, 0, 2])]; + tensor var_4862_end_0 = const()[name = tensor("op_4862_end_0"), val = tensor([1, 1024, 10])]; + tensor var_4862_end_mask_0 = const()[name = tensor("op_4862_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4862 = slice_by_index(begin = var_4862_begin_0, end = var_4862_end_0, end_mask = var_4862_end_mask_0, x = new_x_87)[name = tensor("op_4862")]; + tensor x_567_pad_type_0 = const()[name = tensor("x_567_pad_type_0"), val = tensor("valid")]; + tensor x_567_groups_0 = const()[name = tensor("x_567_groups_0"), val = tensor(1024)]; + tensor x_567_strides_0 = const()[name = tensor("x_567_strides_0"), val = tensor([1])]; + tensor x_567_pad_0 = const()[name = tensor("x_567_pad_0"), val = tensor([0, 0])]; + tensor x_567_dilations_0 = const()[name = tensor("x_567_dilations_0"), val = tensor([1])]; + tensor x_567 = conv(dilations = x_567_dilations_0, groups = x_567_groups_0, pad = x_567_pad_0, pad_type = x_567_pad_type_0, strides = x_567_strides_0, weight = module_layers_21_conv_depthwise_conv_weight_quantized, x = new_x_87)[name = tensor("x_567")]; + tensor input_1147_perm_0 = const()[name = tensor("input_1147_perm_0"), val = tensor([0, 2, 1])]; + tensor x_569_axes_0 = const()[name = tensor("x_569_axes_0"), val = tensor([-1])]; + tensor input_1147 = transpose(perm = input_1147_perm_0, x = x_567)[name = tensor("transpose_167")]; + tensor x_569 = layer_norm(axes = x_569_axes_0, beta = module_layers_21_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_21_conv_batch_norm_weight, x = input_1147)[name = tensor("x_569")]; + tensor input_1149_perm_0 = const()[name = tensor("input_1149_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1149 = transpose(perm = input_1149_perm_0, x = x_569)[name = tensor("transpose_166")]; + tensor input_1151 = silu(x = input_1149)[name = tensor("input_1151")]; + tensor x_571_pad_type_0 = const()[name = tensor("x_571_pad_type_0"), val = tensor("valid")]; + tensor x_571_strides_0 = const()[name = tensor("x_571_strides_0"), val = tensor([1])]; + tensor x_571_pad_0 = const()[name = tensor("x_571_pad_0"), val = tensor([0, 0])]; + tensor x_571_dilations_0 = const()[name = tensor("x_571_dilations_0"), val = tensor([1])]; + tensor x_571_groups_0 = const()[name = tensor("x_571_groups_0"), val = tensor(1)]; + tensor x_571 = conv(dilations = x_571_dilations_0, groups = x_571_groups_0, pad = x_571_pad_0, pad_type = x_571_pad_type_0, strides = x_571_strides_0, weight = module_layers_21_conv_pointwise_conv2_weight_quantized, x = input_1151)[name = tensor("x_571")]; + tensor input_1153_perm_0 = const()[name = tensor("input_1153_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1153 = transpose(perm = input_1153_perm_0, x = x_571)[name = tensor("transpose_165")]; + tensor input_1155 = add(x = input_1139, y = input_1153)[name = tensor("input_1155")]; + tensor input_1157_axes_0 = const()[name = tensor("input_1157_axes_0"), val = tensor([-1])]; + tensor input_1157 = layer_norm(axes = input_1157_axes_0, beta = module_layers_21_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_21_norm_feed_forward2_weight, x = input_1155)[name = tensor("input_1157")]; + tensor input_1159 = linear(bias = linear_1_bias_0, weight = module_layers_21_feed_forward2_linear1_weight_quantized, x = input_1157)[name = tensor("linear_197")]; + tensor input_1161 = silu(x = input_1159)[name = tensor("input_1161")]; + tensor input_1165 = linear(bias = linear_2_bias_0, weight = module_layers_21_feed_forward2_linear2_weight_quantized, x = input_1161)[name = tensor("linear_198")]; + tensor var_4903 = const()[name = tensor("op_4903"), val = tensor(0x1p-1)]; + tensor var_4904 = mul(x = input_1165, y = var_4903)[name = tensor("op_4904")]; + tensor input_1167 = add(x = input_1155, y = var_4904)[name = tensor("input_1167")]; + tensor input_1169_axes_0 = const()[name = tensor("input_1169_axes_0"), val = tensor([-1])]; + tensor input_1169 = layer_norm(axes = input_1169_axes_0, beta = module_layers_21_norm_out_bias, epsilon = var_38, gamma = module_layers_21_norm_out_weight, x = input_1167)[name = tensor("input_1169")]; + tensor cache_89_begin_0 = const()[name = tensor("cache_89_begin_0"), val = tensor([22, 0, 0, 0])]; + tensor cache_89_end_0 = const()[name = tensor("cache_89_end_0"), val = tensor([23, 1, 70, 1024])]; + tensor cache_89_end_mask_0 = const()[name = tensor("cache_89_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_89_squeeze_mask_0 = const()[name = tensor("cache_89_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_89 = slice_by_index(begin = cache_89_begin_0, end = cache_89_end_0, end_mask = cache_89_end_mask_0, squeeze_mask = cache_89_squeeze_mask_0, x = value_3)[name = tensor("cache_89")]; + tensor cache_91_begin_0 = const()[name = tensor("cache_91_begin_0"), val = tensor([22, 0, 0, 0])]; + tensor cache_91_end_0 = const()[name = tensor("cache_91_end_0"), val = tensor([23, 1, 1024, 8])]; + tensor cache_91_end_mask_0 = const()[name = tensor("cache_91_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_91_squeeze_mask_0 = const()[name = tensor("cache_91_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_91 = slice_by_index(begin = cache_91_begin_0, end = cache_91_end_0, end_mask = cache_91_end_mask_0, squeeze_mask = cache_91_squeeze_mask_0, x = value_5)[name = tensor("cache_91")]; + tensor input_1171_axes_0 = const()[name = tensor("input_1171_axes_0"), val = tensor([-1])]; + tensor input_1171 = layer_norm(axes = input_1171_axes_0, beta = module_layers_22_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_22_norm_feed_forward1_weight, x = input_1169)[name = tensor("input_1171")]; + tensor input_1173 = linear(bias = linear_1_bias_0, weight = module_layers_22_feed_forward1_linear1_weight_quantized, x = input_1171)[name = tensor("linear_199")]; + tensor input_1175 = silu(x = input_1173)[name = tensor("input_1175")]; + tensor input_1179 = linear(bias = linear_2_bias_0, weight = module_layers_22_feed_forward1_linear2_weight_quantized, x = input_1175)[name = tensor("linear_200")]; + tensor var_4938 = const()[name = tensor("op_4938"), val = tensor(0x1p-1)]; + tensor var_4939 = mul(x = input_1179, y = var_4938)[name = tensor("op_4939")]; + tensor input_1181 = add(x = input_1169, y = var_4939)[name = tensor("input_1181")]; + tensor key_45_axes_0 = const()[name = tensor("key_45_axes_0"), val = tensor([-1])]; + tensor key_45 = layer_norm(axes = key_45_axes_0, beta = module_layers_22_norm_self_att_bias, epsilon = var_38, gamma = module_layers_22_norm_self_att_weight, x = input_1181)[name = tensor("key_45")]; + tensor input_1183_interleave_0 = const()[name = tensor("input_1183_interleave_0"), val = tensor(false)]; + tensor input_1183 = concat(axis = var_65, interleave = input_1183_interleave_0, values = (cache_89, key_45))[name = tensor("input_1183")]; + tensor var_4961_begin_0 = const()[name = tensor("op_4961_begin_0"), val = tensor([0, 2, 0])]; + tensor var_4961_end_0 = const()[name = tensor("op_4961_end_0"), val = tensor([1, 70, 1024])]; + tensor var_4961_end_mask_0 = const()[name = tensor("op_4961_end_mask_0"), val = tensor([true, true, true])]; + tensor var_4961 = slice_by_index(begin = var_4961_begin_0, end = var_4961_end_0, end_mask = var_4961_end_mask_0, x = cache_89)[name = tensor("op_4961")]; + tensor var_4967_interleave_0 = const()[name = tensor("op_4967_interleave_0"), val = tensor(false)]; + tensor var_4967 = concat(axis = var_65, interleave = var_4967_interleave_0, values = (var_4961, key_45))[name = tensor("op_4967")]; + tensor var_4970 = linear(bias = linear_2_bias_0, weight = module_layers_22_self_attn_linear_q_weight_quantized, x = key_45)[name = tensor("linear_201")]; + tensor var_4971 = const()[name = tensor("op_4971"), val = tensor([1, -1, 8, 128])]; + tensor q_133 = reshape(shape = var_4971, x = var_4970)[name = tensor("q_133")]; + tensor var_4974 = linear(bias = linear_2_bias_0, weight = module_layers_22_self_attn_linear_k_weight_quantized, x = input_1183)[name = tensor("linear_202")]; + tensor var_4975 = const()[name = tensor("op_4975"), val = tensor([1, -1, 8, 128])]; + tensor k_89 = reshape(shape = var_4975, x = var_4974)[name = tensor("k_89")]; + tensor var_4978 = linear(bias = linear_2_bias_0, weight = module_layers_22_self_attn_linear_v_weight_quantized, x = input_1183)[name = tensor("linear_203")]; + tensor var_4979 = const()[name = tensor("op_4979"), val = tensor([1, -1, 8, 128])]; + tensor v_45 = reshape(shape = var_4979, x = var_4978)[name = tensor("v_45")]; + tensor value_53_perm_0 = const()[name = tensor("value_53_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_4991 = add(x = q_133, y = module_layers_22_self_attn_pos_bias_u)[name = tensor("op_4991")]; + tensor var_4993 = add(x = q_133, y = module_layers_22_self_attn_pos_bias_v)[name = tensor("op_4993")]; + tensor q_with_bias_v_45_perm_0 = const()[name = tensor("q_with_bias_v_45_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_4995_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_4995_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590575808))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590722560))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590722304)))]; + tensor x_579_transpose_x_0 = const()[name = tensor("x_579_transpose_x_0"), val = tensor(false)]; + tensor x_579_transpose_y_0 = const()[name = tensor("x_579_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v_45 = transpose(perm = q_with_bias_v_45_perm_0, x = var_4993)[name = tensor("transpose_164")]; + tensor x_579 = matmul(transpose_x = x_579_transpose_x_0, transpose_y = x_579_transpose_y_0, x = q_with_bias_v_45, y = op_4995_quantized)[name = tensor("x_579")]; + tensor const_365 = const()[name = tensor("const_365"), val = tensor(0x0p+0)]; + tensor x_581_pad_0 = const()[name = tensor("x_581_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_581_mode_0 = const()[name = tensor("x_581_mode_0"), val = tensor("constant")]; + tensor x_581 = pad(constant_val = const_365, mode = x_581_mode_0, pad = x_581_pad_0, x = x_579)[name = tensor("x_581")]; + tensor var_5003 = const()[name = tensor("op_5003"), val = tensor([1, 8, -1, 2])]; + tensor x_583 = reshape(shape = var_5003, x = x_581)[name = tensor("x_583")]; + tensor var_5007_begin_0 = const()[name = tensor("op_5007_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_5007_end_0 = const()[name = tensor("op_5007_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_5007_end_mask_0 = const()[name = tensor("op_5007_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_5007 = slice_by_index(begin = var_5007_begin_0, end = var_5007_end_0, end_mask = var_5007_end_mask_0, x = x_583)[name = tensor("op_5007")]; + tensor var_5008 = const()[name = tensor("op_5008"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_89 = reshape(shape = var_5008, x = var_5007)[name = tensor("matrix_bd_89")]; + tensor matrix_ac_45_transpose_x_0 = const()[name = tensor("matrix_ac_45_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_45_transpose_y_0 = const()[name = tensor("matrix_ac_45_transpose_y_0"), val = tensor(false)]; + tensor transpose_140_perm_0 = const()[name = tensor("transpose_140_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_141_perm_0 = const()[name = tensor("transpose_141_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_141 = transpose(perm = transpose_141_perm_0, x = k_89)[name = tensor("transpose_162")]; + tensor transpose_140 = transpose(perm = transpose_140_perm_0, x = var_4991)[name = tensor("transpose_163")]; + tensor matrix_ac_45 = matmul(transpose_x = matrix_ac_45_transpose_x_0, transpose_y = matrix_ac_45_transpose_y_0, x = transpose_140, y = transpose_141)[name = tensor("matrix_ac_45")]; + tensor matrix_bd_91_begin_0 = const()[name = tensor("matrix_bd_91_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_91_end_0 = const()[name = tensor("matrix_bd_91_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_91_end_mask_0 = const()[name = tensor("matrix_bd_91_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd_91 = slice_by_index(begin = matrix_bd_91_begin_0, end = matrix_bd_91_end_0, end_mask = matrix_bd_91_end_mask_0, x = matrix_bd_89)[name = tensor("matrix_bd_91")]; + tensor var_5017 = add(x = matrix_ac_45, y = matrix_bd_91)[name = tensor("op_5017")]; + tensor _inversed_scores_89_y_0 = const()[name = tensor("_inversed_scores_89_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_89 = mul(x = var_5017, y = _inversed_scores_89_y_0)[name = tensor("_inversed_scores_89")]; + tensor scores_91 = select(a = var_41, b = _inversed_scores_89, cond = mask_11)[name = tensor("scores_91")]; + tensor var_5023 = softmax(axis = var_56, x = scores_91)[name = tensor("op_5023")]; + tensor input_1185 = select(a = var_40, b = var_5023, cond = mask_11)[name = tensor("input_1185")]; + tensor x_585_transpose_x_0 = const()[name = tensor("x_585_transpose_x_0"), val = tensor(false)]; + tensor x_585_transpose_y_0 = const()[name = tensor("x_585_transpose_y_0"), val = tensor(false)]; + tensor value_53 = transpose(perm = value_53_perm_0, x = v_45)[name = tensor("transpose_161")]; + tensor x_585 = matmul(transpose_x = x_585_transpose_x_0, transpose_y = x_585_transpose_y_0, x = input_1185, y = value_53)[name = tensor("x_585")]; + tensor var_5027_perm_0 = const()[name = tensor("op_5027_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_5028 = const()[name = tensor("op_5028"), val = tensor([1, -1, 1024])]; + tensor var_5027 = transpose(perm = var_5027_perm_0, x = x_585)[name = tensor("transpose_160")]; + tensor input_1187 = reshape(shape = var_5028, x = var_5027)[name = tensor("input_1187")]; + tensor input_1189 = linear(bias = linear_2_bias_0, weight = module_layers_22_self_attn_linear_out_weight_quantized, x = input_1187)[name = tensor("linear_205")]; + tensor input_1191 = add(x = input_1181, y = input_1189)[name = tensor("input_1191")]; + tensor x_589_axes_0 = const()[name = tensor("x_589_axes_0"), val = tensor([-1])]; + tensor x_589 = layer_norm(axes = x_589_axes_0, beta = module_layers_22_norm_conv_bias, epsilon = var_38, gamma = module_layers_22_norm_conv_weight, x = input_1191)[name = tensor("x_589")]; + tensor input_1193_perm_0 = const()[name = tensor("input_1193_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1195_pad_type_0 = const()[name = tensor("input_1195_pad_type_0"), val = tensor("valid")]; + tensor input_1195_strides_0 = const()[name = tensor("input_1195_strides_0"), val = tensor([1])]; + tensor input_1195_pad_0 = const()[name = tensor("input_1195_pad_0"), val = tensor([0, 0])]; + tensor input_1195_dilations_0 = const()[name = tensor("input_1195_dilations_0"), val = tensor([1])]; + tensor input_1195_groups_0 = const()[name = tensor("input_1195_groups_0"), val = tensor(1)]; + tensor input_1193 = transpose(perm = input_1193_perm_0, x = x_589)[name = tensor("transpose_159")]; + tensor input_1195 = conv(dilations = input_1195_dilations_0, groups = input_1195_groups_0, pad = input_1195_pad_0, pad_type = input_1195_pad_type_0, strides = input_1195_strides_0, weight = module_layers_22_conv_pointwise_conv1_weight_quantized, x = input_1193)[name = tensor("input_1195")]; + tensor x_591_split_num_splits_0 = const()[name = tensor("x_591_split_num_splits_0"), val = tensor(2)]; + tensor x_591_split_axis_0 = const()[name = tensor("x_591_split_axis_0"), val = tensor(1)]; + tensor x_591_split_0, tensor x_591_split_1 = split(axis = x_591_split_axis_0, num_splits = x_591_split_num_splits_0, x = input_1195)[name = tensor("x_591_split")]; + tensor x_591_split_1_sigmoid = sigmoid(x = x_591_split_1)[name = tensor("x_591_split_1_sigmoid")]; + tensor x_591 = mul(x = x_591_split_0, y = x_591_split_1_sigmoid)[name = tensor("x_591")]; + tensor input_1197 = select(a = var_40, b = x_591, cond = var_565)[name = tensor("input_1197")]; + tensor new_x_91_interleave_0 = const()[name = tensor("new_x_91_interleave_0"), val = tensor(false)]; + tensor new_x_91 = concat(axis = var_56, interleave = new_x_91_interleave_0, values = (cache_91, input_1197))[name = tensor("new_x_91")]; + tensor var_5066_begin_0 = const()[name = tensor("op_5066_begin_0"), val = tensor([0, 0, 2])]; + tensor var_5066_end_0 = const()[name = tensor("op_5066_end_0"), val = tensor([1, 1024, 10])]; + tensor var_5066_end_mask_0 = const()[name = tensor("op_5066_end_mask_0"), val = tensor([true, true, true])]; + tensor var_5066 = slice_by_index(begin = var_5066_begin_0, end = var_5066_end_0, end_mask = var_5066_end_mask_0, x = new_x_91)[name = tensor("op_5066")]; + tensor x_593_pad_type_0 = const()[name = tensor("x_593_pad_type_0"), val = tensor("valid")]; + tensor x_593_groups_0 = const()[name = tensor("x_593_groups_0"), val = tensor(1024)]; + tensor x_593_strides_0 = const()[name = tensor("x_593_strides_0"), val = tensor([1])]; + tensor x_593_pad_0 = const()[name = tensor("x_593_pad_0"), val = tensor([0, 0])]; + tensor x_593_dilations_0 = const()[name = tensor("x_593_dilations_0"), val = tensor([1])]; + tensor x_593 = conv(dilations = x_593_dilations_0, groups = x_593_groups_0, pad = x_593_pad_0, pad_type = x_593_pad_type_0, strides = x_593_strides_0, weight = module_layers_22_conv_depthwise_conv_weight_quantized, x = new_x_91)[name = tensor("x_593")]; + tensor input_1199_perm_0 = const()[name = tensor("input_1199_perm_0"), val = tensor([0, 2, 1])]; + tensor x_595_axes_0 = const()[name = tensor("x_595_axes_0"), val = tensor([-1])]; + tensor input_1199 = transpose(perm = input_1199_perm_0, x = x_593)[name = tensor("transpose_158")]; + tensor x_595 = layer_norm(axes = x_595_axes_0, beta = module_layers_22_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_22_conv_batch_norm_weight, x = input_1199)[name = tensor("x_595")]; + tensor input_1201_perm_0 = const()[name = tensor("input_1201_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1201 = transpose(perm = input_1201_perm_0, x = x_595)[name = tensor("transpose_157")]; + tensor input_1203 = silu(x = input_1201)[name = tensor("input_1203")]; + tensor x_597_pad_type_0 = const()[name = tensor("x_597_pad_type_0"), val = tensor("valid")]; + tensor x_597_strides_0 = const()[name = tensor("x_597_strides_0"), val = tensor([1])]; + tensor x_597_pad_0 = const()[name = tensor("x_597_pad_0"), val = tensor([0, 0])]; + tensor x_597_dilations_0 = const()[name = tensor("x_597_dilations_0"), val = tensor([1])]; + tensor x_597_groups_0 = const()[name = tensor("x_597_groups_0"), val = tensor(1)]; + tensor x_597 = conv(dilations = x_597_dilations_0, groups = x_597_groups_0, pad = x_597_pad_0, pad_type = x_597_pad_type_0, strides = x_597_strides_0, weight = module_layers_22_conv_pointwise_conv2_weight_quantized, x = input_1203)[name = tensor("x_597")]; + tensor input_1205_perm_0 = const()[name = tensor("input_1205_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1205 = transpose(perm = input_1205_perm_0, x = x_597)[name = tensor("transpose_156")]; + tensor input_1207 = add(x = input_1191, y = input_1205)[name = tensor("input_1207")]; + tensor input_1209_axes_0 = const()[name = tensor("input_1209_axes_0"), val = tensor([-1])]; + tensor input_1209 = layer_norm(axes = input_1209_axes_0, beta = module_layers_22_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_22_norm_feed_forward2_weight, x = input_1207)[name = tensor("input_1209")]; + tensor input_1211 = linear(bias = linear_1_bias_0, weight = module_layers_22_feed_forward2_linear1_weight_quantized, x = input_1209)[name = tensor("linear_206")]; + tensor input_1213 = silu(x = input_1211)[name = tensor("input_1213")]; + tensor input_1217 = linear(bias = linear_2_bias_0, weight = module_layers_22_feed_forward2_linear2_weight_quantized, x = input_1213)[name = tensor("linear_207")]; + tensor var_5107 = const()[name = tensor("op_5107"), val = tensor(0x1p-1)]; + tensor var_5108 = mul(x = input_1217, y = var_5107)[name = tensor("op_5108")]; + tensor input_1219 = add(x = input_1207, y = var_5108)[name = tensor("input_1219")]; + tensor input_1221_axes_0 = const()[name = tensor("input_1221_axes_0"), val = tensor([-1])]; + tensor input_1221 = layer_norm(axes = input_1221_axes_0, beta = module_layers_22_norm_out_bias, epsilon = var_38, gamma = module_layers_22_norm_out_weight, x = input_1219)[name = tensor("input_1221")]; + tensor cache_93_begin_0 = const()[name = tensor("cache_93_begin_0"), val = tensor([23, 0, 0, 0])]; + tensor cache_93_end_0 = const()[name = tensor("cache_93_end_0"), val = tensor([24, 1, 70, 1024])]; + tensor cache_93_end_mask_0 = const()[name = tensor("cache_93_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_93_squeeze_mask_0 = const()[name = tensor("cache_93_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache_93 = slice_by_index(begin = cache_93_begin_0, end = cache_93_end_0, end_mask = cache_93_end_mask_0, squeeze_mask = cache_93_squeeze_mask_0, x = value_3)[name = tensor("cache_93")]; + tensor cache_begin_0 = const()[name = tensor("cache_begin_0"), val = tensor([23, 0, 0, 0])]; + tensor cache_end_0 = const()[name = tensor("cache_end_0"), val = tensor([24, 1, 1024, 8])]; + tensor cache_end_mask_0 = const()[name = tensor("cache_end_mask_0"), val = tensor([false, true, true, true])]; + tensor cache_squeeze_mask_0 = const()[name = tensor("cache_squeeze_mask_0"), val = tensor([true, false, false, false])]; + tensor cache = slice_by_index(begin = cache_begin_0, end = cache_end_0, end_mask = cache_end_mask_0, squeeze_mask = cache_squeeze_mask_0, x = value_5)[name = tensor("cache")]; + tensor input_1223_axes_0 = const()[name = tensor("input_1223_axes_0"), val = tensor([-1])]; + tensor input_1223 = layer_norm(axes = input_1223_axes_0, beta = module_layers_23_norm_feed_forward1_bias, epsilon = var_38, gamma = module_layers_23_norm_feed_forward1_weight, x = input_1221)[name = tensor("input_1223")]; + tensor input_1225 = linear(bias = linear_1_bias_0, weight = module_layers_23_feed_forward1_linear1_weight_quantized, x = input_1223)[name = tensor("linear_208")]; + tensor input_1227 = silu(x = input_1225)[name = tensor("input_1227")]; + tensor input_1231 = linear(bias = linear_2_bias_0, weight = module_layers_23_feed_forward1_linear2_weight_quantized, x = input_1227)[name = tensor("linear_209")]; + tensor var_5142 = const()[name = tensor("op_5142"), val = tensor(0x1p-1)]; + tensor var_5143 = mul(x = input_1231, y = var_5142)[name = tensor("op_5143")]; + tensor input_1233 = add(x = input_1221, y = var_5143)[name = tensor("input_1233")]; + tensor key_axes_0 = const()[name = tensor("key_axes_0"), val = tensor([-1])]; + tensor key = layer_norm(axes = key_axes_0, beta = module_layers_23_norm_self_att_bias, epsilon = var_38, gamma = module_layers_23_norm_self_att_weight, x = input_1233)[name = tensor("key")]; + tensor input_1235_interleave_0 = const()[name = tensor("input_1235_interleave_0"), val = tensor(false)]; + tensor input_1235 = concat(axis = var_65, interleave = input_1235_interleave_0, values = (cache_93, key))[name = tensor("input_1235")]; + tensor var_5165_begin_0 = const()[name = tensor("op_5165_begin_0"), val = tensor([0, 2, 0])]; + tensor var_5165_end_0 = const()[name = tensor("op_5165_end_0"), val = tensor([1, 70, 1024])]; + tensor var_5165_end_mask_0 = const()[name = tensor("op_5165_end_mask_0"), val = tensor([true, true, true])]; + tensor var_5165 = slice_by_index(begin = var_5165_begin_0, end = var_5165_end_0, end_mask = var_5165_end_mask_0, x = cache_93)[name = tensor("op_5165")]; + tensor cache_last_channel_cur_interleave_0 = const()[name = tensor("cache_last_channel_cur_interleave_0"), val = tensor(false)]; + tensor cache_last_channel_cur = concat(axis = var_65, interleave = cache_last_channel_cur_interleave_0, values = (var_5165, key))[name = tensor("cache_last_channel_cur")]; + tensor var_5174 = linear(bias = linear_2_bias_0, weight = module_layers_23_self_attn_linear_q_weight_quantized, x = key)[name = tensor("linear_210")]; + tensor var_5175 = const()[name = tensor("op_5175"), val = tensor([1, -1, 8, 128])]; + tensor q_139 = reshape(shape = var_5175, x = var_5174)[name = tensor("q_139")]; + tensor var_5178 = linear(bias = linear_2_bias_0, weight = module_layers_23_self_attn_linear_k_weight_quantized, x = input_1235)[name = tensor("linear_211")]; + tensor var_5179 = const()[name = tensor("op_5179"), val = tensor([1, -1, 8, 128])]; + tensor k_93 = reshape(shape = var_5179, x = var_5178)[name = tensor("k_93")]; + tensor var_5182 = linear(bias = linear_2_bias_0, weight = module_layers_23_self_attn_linear_v_weight_quantized, x = input_1235)[name = tensor("linear_212")]; + tensor var_5183 = const()[name = tensor("op_5183"), val = tensor([1, -1, 8, 128])]; + tensor v = reshape(shape = var_5183, x = var_5182)[name = tensor("v")]; + tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor var_5195 = add(x = q_139, y = module_layers_23_self_attn_pos_bias_u)[name = tensor("op_5195")]; + tensor var_5197 = add(x = q_139, y = module_layers_23_self_attn_pos_bias_v)[name = tensor("op_5197")]; + tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor op_5199_quantized = constexpr_affine_dequantize()[axis = tensor(3), name = tensor("op_5199_quantized"), quantized_data = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590723200))), scale = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590869952))), zero_point = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(590869696)))]; + tensor x_605_transpose_x_0 = const()[name = tensor("x_605_transpose_x_0"), val = tensor(false)]; + tensor x_605_transpose_y_0 = const()[name = tensor("x_605_transpose_y_0"), val = tensor(false)]; + tensor q_with_bias_v = transpose(perm = q_with_bias_v_perm_0, x = var_5197)[name = tensor("transpose_155")]; + tensor x_605 = matmul(transpose_x = x_605_transpose_x_0, transpose_y = x_605_transpose_y_0, x = q_with_bias_v, y = op_5199_quantized)[name = tensor("x_605")]; + tensor const_378 = const()[name = tensor("const_378"), val = tensor(0x0p+0)]; + tensor x_607_pad_0 = const()[name = tensor("x_607_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; + tensor x_607_mode_0 = const()[name = tensor("x_607_mode_0"), val = tensor("constant")]; + tensor x_607 = pad(constant_val = const_378, mode = x_607_mode_0, pad = x_607_pad_0, x = x_605)[name = tensor("x_607")]; + tensor var_5207 = const()[name = tensor("op_5207"), val = tensor([1, 8, -1, 2])]; + tensor x_609 = reshape(shape = var_5207, x = x_607)[name = tensor("x_609")]; + tensor var_5211_begin_0 = const()[name = tensor("op_5211_begin_0"), val = tensor([0, 0, 1, 0])]; + tensor var_5211_end_0 = const()[name = tensor("op_5211_end_0"), val = tensor([1, 8, 144, 2])]; + tensor var_5211_end_mask_0 = const()[name = tensor("op_5211_end_mask_0"), val = tensor([true, true, true, true])]; + tensor var_5211 = slice_by_index(begin = var_5211_begin_0, end = var_5211_end_0, end_mask = var_5211_end_mask_0, x = x_609)[name = tensor("op_5211")]; + tensor var_5212 = const()[name = tensor("op_5212"), val = tensor([1, 8, 2, 143])]; + tensor matrix_bd_93 = reshape(shape = var_5212, x = var_5211)[name = tensor("matrix_bd_93")]; + tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; + tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; + tensor transpose_142_perm_0 = const()[name = tensor("transpose_142_perm_0"), val = tensor([0, 2, -3, -1])]; + tensor transpose_143_perm_0 = const()[name = tensor("transpose_143_perm_0"), val = tensor([0, 2, -1, -3])]; + tensor transpose_143 = transpose(perm = transpose_143_perm_0, x = k_93)[name = tensor("transpose_153")]; + tensor transpose_142 = transpose(perm = transpose_142_perm_0, x = var_5195)[name = tensor("transpose_154")]; + tensor matrix_ac = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_142, y = transpose_143)[name = tensor("matrix_ac")]; + tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; + tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 2, 72])]; + tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; + tensor matrix_bd = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_93)[name = tensor("matrix_bd")]; + tensor var_5221 = add(x = matrix_ac, y = matrix_bd)[name = tensor("op_5221")]; + tensor _inversed_scores_93_y_0 = const()[name = tensor("_inversed_scores_93_y_0"), val = tensor(0x1.6a09e6p-4)]; + tensor _inversed_scores_93 = mul(x = var_5221, y = _inversed_scores_93_y_0)[name = tensor("_inversed_scores_93")]; + tensor scores = select(a = var_41, b = _inversed_scores_93, cond = mask_11)[name = tensor("scores")]; + tensor var_5227 = softmax(axis = var_56, x = scores)[name = tensor("op_5227")]; + tensor input_1237 = select(a = var_40, b = var_5227, cond = mask_11)[name = tensor("input_1237")]; + tensor x_611_transpose_x_0 = const()[name = tensor("x_611_transpose_x_0"), val = tensor(false)]; + tensor x_611_transpose_y_0 = const()[name = tensor("x_611_transpose_y_0"), val = tensor(false)]; + tensor value = transpose(perm = value_perm_0, x = v)[name = tensor("transpose_152")]; + tensor x_611 = matmul(transpose_x = x_611_transpose_x_0, transpose_y = x_611_transpose_y_0, x = input_1237, y = value)[name = tensor("x_611")]; + tensor var_5231_perm_0 = const()[name = tensor("op_5231_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_5232 = const()[name = tensor("op_5232"), val = tensor([1, -1, 1024])]; + tensor var_5231 = transpose(perm = var_5231_perm_0, x = x_611)[name = tensor("transpose_151")]; + tensor input_1239 = reshape(shape = var_5232, x = var_5231)[name = tensor("input_1239")]; + tensor input_1241 = linear(bias = linear_2_bias_0, weight = module_layers_23_self_attn_linear_out_weight_quantized, x = input_1239)[name = tensor("linear_214")]; + tensor input_1243 = add(x = input_1233, y = input_1241)[name = tensor("input_1243")]; + tensor x_615_axes_0 = const()[name = tensor("x_615_axes_0"), val = tensor([-1])]; + tensor x_615 = layer_norm(axes = x_615_axes_0, beta = module_layers_23_norm_conv_bias, epsilon = var_38, gamma = module_layers_23_norm_conv_weight, x = input_1243)[name = tensor("x_615")]; + tensor input_1245_perm_0 = const()[name = tensor("input_1245_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1247_pad_type_0 = const()[name = tensor("input_1247_pad_type_0"), val = tensor("valid")]; + tensor input_1247_strides_0 = const()[name = tensor("input_1247_strides_0"), val = tensor([1])]; + tensor input_1247_pad_0 = const()[name = tensor("input_1247_pad_0"), val = tensor([0, 0])]; + tensor input_1247_dilations_0 = const()[name = tensor("input_1247_dilations_0"), val = tensor([1])]; + tensor input_1247_groups_0 = const()[name = tensor("input_1247_groups_0"), val = tensor(1)]; + tensor input_1245 = transpose(perm = input_1245_perm_0, x = x_615)[name = tensor("transpose_150")]; + tensor input_1247 = conv(dilations = input_1247_dilations_0, groups = input_1247_groups_0, pad = input_1247_pad_0, pad_type = input_1247_pad_type_0, strides = input_1247_strides_0, weight = module_layers_23_conv_pointwise_conv1_weight_quantized, x = input_1245)[name = tensor("input_1247")]; + tensor x_617_split_num_splits_0 = const()[name = tensor("x_617_split_num_splits_0"), val = tensor(2)]; + tensor x_617_split_axis_0 = const()[name = tensor("x_617_split_axis_0"), val = tensor(1)]; + tensor x_617_split_0, tensor x_617_split_1 = split(axis = x_617_split_axis_0, num_splits = x_617_split_num_splits_0, x = input_1247)[name = tensor("x_617_split")]; + tensor x_617_split_1_sigmoid = sigmoid(x = x_617_split_1)[name = tensor("x_617_split_1_sigmoid")]; + tensor x_617 = mul(x = x_617_split_0, y = x_617_split_1_sigmoid)[name = tensor("x_617")]; + tensor input_1249 = select(a = var_40, b = x_617, cond = var_565)[name = tensor("input_1249")]; + tensor new_x_interleave_0 = const()[name = tensor("new_x_interleave_0"), val = tensor(false)]; + tensor new_x = concat(axis = var_56, interleave = new_x_interleave_0, values = (cache, input_1249))[name = tensor("new_x")]; + tensor cache_last_time_cur_begin_0 = const()[name = tensor("cache_last_time_cur_begin_0"), val = tensor([0, 0, 2])]; + tensor cache_last_time_cur_end_0 = const()[name = tensor("cache_last_time_cur_end_0"), val = tensor([1, 1024, 10])]; + tensor cache_last_time_cur_end_mask_0 = const()[name = tensor("cache_last_time_cur_end_mask_0"), val = tensor([true, true, true])]; + tensor cache_last_time_cur = slice_by_index(begin = cache_last_time_cur_begin_0, end = cache_last_time_cur_end_0, end_mask = cache_last_time_cur_end_mask_0, x = new_x)[name = tensor("cache_last_time_cur")]; + tensor x_619_pad_type_0 = const()[name = tensor("x_619_pad_type_0"), val = tensor("valid")]; + tensor x_619_groups_0 = const()[name = tensor("x_619_groups_0"), val = tensor(1024)]; + tensor x_619_strides_0 = const()[name = tensor("x_619_strides_0"), val = tensor([1])]; + tensor x_619_pad_0 = const()[name = tensor("x_619_pad_0"), val = tensor([0, 0])]; + tensor x_619_dilations_0 = const()[name = tensor("x_619_dilations_0"), val = tensor([1])]; + tensor x_619 = conv(dilations = x_619_dilations_0, groups = x_619_groups_0, pad = x_619_pad_0, pad_type = x_619_pad_type_0, strides = x_619_strides_0, weight = module_layers_23_conv_depthwise_conv_weight_quantized, x = new_x)[name = tensor("x_619")]; + tensor input_1251_perm_0 = const()[name = tensor("input_1251_perm_0"), val = tensor([0, 2, 1])]; + tensor x_621_axes_0 = const()[name = tensor("x_621_axes_0"), val = tensor([-1])]; + tensor input_1251 = transpose(perm = input_1251_perm_0, x = x_619)[name = tensor("transpose_149")]; + tensor x_621 = layer_norm(axes = x_621_axes_0, beta = module_layers_23_conv_batch_norm_bias, epsilon = var_38, gamma = module_layers_23_conv_batch_norm_weight, x = input_1251)[name = tensor("x_621")]; + tensor input_1253_perm_0 = const()[name = tensor("input_1253_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1253 = transpose(perm = input_1253_perm_0, x = x_621)[name = tensor("transpose_148")]; + tensor input_1255 = silu(x = input_1253)[name = tensor("input_1255")]; + tensor x_623_pad_type_0 = const()[name = tensor("x_623_pad_type_0"), val = tensor("valid")]; + tensor x_623_strides_0 = const()[name = tensor("x_623_strides_0"), val = tensor([1])]; + tensor x_623_pad_0 = const()[name = tensor("x_623_pad_0"), val = tensor([0, 0])]; + tensor x_623_dilations_0 = const()[name = tensor("x_623_dilations_0"), val = tensor([1])]; + tensor x_623_groups_0 = const()[name = tensor("x_623_groups_0"), val = tensor(1)]; + tensor x_623 = conv(dilations = x_623_dilations_0, groups = x_623_groups_0, pad = x_623_pad_0, pad_type = x_623_pad_type_0, strides = x_623_strides_0, weight = module_layers_23_conv_pointwise_conv2_weight_quantized, x = input_1255)[name = tensor("x_623")]; + tensor input_1257_perm_0 = const()[name = tensor("input_1257_perm_0"), val = tensor([0, 2, 1])]; + tensor input_1257 = transpose(perm = input_1257_perm_0, x = x_623)[name = tensor("transpose_147")]; + tensor input_1259 = add(x = input_1243, y = input_1257)[name = tensor("input_1259")]; + tensor input_1261_axes_0 = const()[name = tensor("input_1261_axes_0"), val = tensor([-1])]; + tensor input_1261 = layer_norm(axes = input_1261_axes_0, beta = module_layers_23_norm_feed_forward2_bias, epsilon = var_38, gamma = module_layers_23_norm_feed_forward2_weight, x = input_1259)[name = tensor("input_1261")]; + tensor input_1263 = linear(bias = linear_1_bias_0, weight = module_layers_23_feed_forward2_linear1_weight_quantized, x = input_1261)[name = tensor("linear_215")]; + tensor input_1265 = silu(x = input_1263)[name = tensor("input_1265")]; + tensor input_1269 = linear(bias = linear_2_bias_0, weight = module_layers_23_feed_forward2_linear2_weight_quantized, x = input_1265)[name = tensor("linear_216")]; + tensor var_5311 = const()[name = tensor("op_5311"), val = tensor(0x1p-1)]; + tensor var_5312 = mul(x = input_1269, y = var_5311)[name = tensor("op_5312")]; + tensor input = add(x = input_1259, y = var_5312)[name = tensor("input")]; + tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; + tensor audio_signal = layer_norm(axes = audio_signal_axes_0, beta = module_layers_23_norm_out_bias, epsilon = var_38, gamma = module_layers_23_norm_out_weight, x = input)[name = tensor("audio_signal")]; + tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; + tensor obj_5_axis_0 = const()[name = tensor("obj_5_axis_0"), val = tensor(0)]; + tensor obj_5 = stack(axis = obj_5_axis_0, values = (var_479, var_683, var_887, var_1091, var_1295, var_1499, var_1703, var_1907, var_2111, var_2315, var_2519, var_2723, var_2927, var_3131, var_3335, var_3539, var_3743, var_3947, var_4151, var_4355, var_4559, var_4763, var_4967, cache_last_channel_cur))[name = tensor("obj_5")]; + tensor obj_7_axis_0 = const()[name = tensor("obj_7_axis_0"), val = tensor(0)]; + tensor obj_7 = stack(axis = obj_7_axis_0, values = (var_578, var_782, var_986, var_1190, var_1394, var_1598, var_1802, var_2006, var_2210, var_2414, var_2618, var_2822, var_3026, var_3230, var_3434, var_3638, var_3842, var_4046, var_4250, var_4454, var_4658, var_4862, var_5066, cache_last_time_cur))[name = tensor("obj_7")]; + tensor var_5328 = add(x = cache_len, y = max_audio_length_1)[name = tensor("op_5328")]; + tensor const_384 = const()[name = tensor("const_384"), val = tensor(-0x1.fffffep+127)]; + tensor var_5328_promoted_dtype_0 = const()[name = tensor("op_5328_promoted_dtype_0"), val = tensor("fp32")]; + tensor var_45_promoted = const()[name = tensor("op_45_promoted"), val = tensor(0x1.18p+6)]; + tensor var_5328_promoted = cast(dtype = var_5328_promoted_dtype_0, x = var_5328)[name = tensor("cast_2")]; + tensor clip_1 = clip(alpha = const_384, beta = var_45_promoted, x = var_5328_promoted)[name = tensor("clip_1")]; + tensor var_5338_perm_0 = const()[name = tensor("op_5338_perm_0"), val = tensor([1, 0, 2, 3])]; + tensor var_5341_perm_0 = const()[name = tensor("op_5341_perm_0"), val = tensor([1, 0, 2, 3])]; + tensor var_5346_dtype_0 = const()[name = tensor("op_5346_dtype_0"), val = tensor("int32")]; + tensor var_5351_dtype_0 = const()[name = tensor("op_5351_dtype_0"), val = tensor("int32")]; + tensor encoded = transpose(perm = obj_1_perm_0, x = audio_signal)[name = tensor("transpose_144")]; + tensor cache_channel_out = transpose(perm = var_5338_perm_0, x = obj_5)[name = tensor("transpose_145")]; + tensor cache_time_out = transpose(perm = var_5341_perm_0, x = obj_7)[name = tensor("transpose_146")]; + tensor encoded_length = cast(dtype = var_5346_dtype_0, x = clip_0)[name = tensor("cast_0")]; + tensor cache_len_out = cast(dtype = var_5351_dtype_0, x = clip_1)[name = tensor("cast_1")]; + } -> (encoded, encoded_length, cache_channel_out, cache_time_out, cache_len_out); +} \ No newline at end of file