diff --git "a/SplitEinsum-Resources/TextEncoder2.mlmodelc/model.mil" "b/SplitEinsum-Resources/TextEncoder2.mlmodelc/model.mil" new file mode 100644--- /dev/null +++ "b/SplitEinsum-Resources/TextEncoder2.mlmodelc/model.mil" @@ -0,0 +1,2211 @@ +program(1.0) +[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3304.5.2"}, {"coremlc-version", "3304.6.2"}, {"coremltools-component-torch", "2.2.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.0b1"}})] +{ + func main(tensor input_ids) { + tensor var_5 = const()[name = tensor("op_5"), val = tensor(-1)]; + tensor var_6 = const()[name = tensor("op_6"), val = tensor(false)]; + tensor cast_1_dtype_0 = const()[name = tensor("cast_1_dtype_0"), val = tensor("int32")]; + tensor inputs_embeds_axis_0 = const()[name = tensor("inputs_embeds_axis_0"), val = tensor(0)]; + tensor inputs_embeds_batch_dims_0 = const()[name = tensor("inputs_embeds_batch_dims_0"), val = tensor(0)]; + tensor text_encoder_text_model_embeddings_token_embedding_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_embeddings_token_embedding_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; + tensor cast_1 = cast(dtype = cast_1_dtype_0, x = input_ids)[name = tensor("cast_329")]; + tensor inputs_embeds_cast_fp16 = gather(axis = inputs_embeds_axis_0, batch_dims = inputs_embeds_batch_dims_0, indices = cast_1, x = text_encoder_text_model_embeddings_token_embedding_weight_to_fp16)[name = tensor("inputs_embeds_cast_fp16")]; + tensor position_embeddings_to_fp16 = const()[name = tensor("position_embeddings_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126484608)))]; + tensor input_3_cast_fp16 = add(x = inputs_embeds_cast_fp16, y = position_embeddings_to_fp16)[name = tensor("input_3_cast_fp16")]; + tensor hidden_states_1_axes_0 = const()[name = tensor("hidden_states_1_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126681792)))]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126684416)))]; + tensor var_15_to_fp16 = const()[name = tensor("op_15_to_fp16"), val = tensor(0x1.5p-17)]; + tensor hidden_states_1_cast_fp16 = layer_norm(axes = hidden_states_1_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16, x = input_3_cast_fp16)[name = tensor("hidden_states_1_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126687040)))]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129963904)))]; + tensor linear_0_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor("linear_0_cast_fp16")]; + tensor var_149_to_fp16 = const()[name = tensor("op_149_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_5_cast_fp16 = mul(x = linear_0_cast_fp16, y = var_149_to_fp16)[name = tensor("tensor_5_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129966528)))]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133243392)))]; + tensor linear_1_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor("linear_1_cast_fp16")]; + tensor var_154 = const()[name = tensor("op_154"), val = tensor([1, -1, 20, 64])]; + tensor var_155_cast_fp16 = reshape(shape = var_154, x = linear_1_cast_fp16)[name = tensor("op_155_cast_fp16")]; + tensor var_156_perm_0 = const()[name = tensor("op_156_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133246016)))]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136522880)))]; + tensor linear_2_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor("linear_2_cast_fp16")]; + tensor var_161 = const()[name = tensor("op_161"), val = tensor([1, -1, 20, 64])]; + tensor var_162_cast_fp16 = reshape(shape = var_161, x = linear_2_cast_fp16)[name = tensor("op_162_cast_fp16")]; + tensor var_163_perm_0 = const()[name = tensor("op_163_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_170 = const()[name = tensor("op_170"), val = tensor([1, 77, 20, 64])]; + tensor var_171_cast_fp16 = reshape(shape = var_170, x = tensor_5_cast_fp16)[name = tensor("op_171_cast_fp16")]; + tensor var_172_perm_0 = const()[name = tensor("op_172_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_174 = const()[name = tensor("op_174"), val = tensor([20, -1, 64])]; + tensor var_172_cast_fp16 = transpose(perm = var_172_perm_0, x = var_171_cast_fp16)[name = tensor("transpose_126")]; + tensor query_states_1_cast_fp16 = reshape(shape = var_174, x = var_172_cast_fp16)[name = tensor("query_states_1_cast_fp16")]; + tensor var_176 = const()[name = tensor("op_176"), val = tensor([20, -1, 64])]; + tensor var_156_cast_fp16 = transpose(perm = var_156_perm_0, x = var_155_cast_fp16)[name = tensor("transpose_128")]; + tensor key_states_3_cast_fp16 = reshape(shape = var_176, x = var_156_cast_fp16)[name = tensor("key_states_3_cast_fp16")]; + tensor var_178 = const()[name = tensor("op_178"), val = tensor([20, -1, 64])]; + tensor var_163_cast_fp16 = transpose(perm = var_163_perm_0, x = var_162_cast_fp16)[name = tensor("transpose_127")]; + tensor value_states_3_cast_fp16 = reshape(shape = var_178, x = var_163_cast_fp16)[name = tensor("value_states_3_cast_fp16")]; + tensor attn_weights_1_transpose_x_1 = const()[name = tensor("attn_weights_1_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_1_transpose_y_1 = const()[name = tensor("attn_weights_1_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_1_cast_fp16 = matmul(transpose_x = attn_weights_1_transpose_x_1, transpose_y = attn_weights_1_transpose_y_1, x = query_states_1_cast_fp16, y = key_states_3_cast_fp16)[name = tensor("attn_weights_1_cast_fp16")]; + tensor var_183 = const()[name = tensor("op_183"), val = tensor([1, 20, 77, 77])]; + tensor var_184_cast_fp16 = reshape(shape = var_183, x = attn_weights_1_cast_fp16)[name = tensor("op_184_cast_fp16")]; + tensor var_58_to_fp16 = const()[name = tensor("op_58_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136525504)))]; + tensor attn_weights_3_cast_fp16 = add(x = var_184_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_3_cast_fp16")]; + tensor var_189 = const()[name = tensor("op_189"), val = tensor([20, 77, 77])]; + tensor input_5_cast_fp16 = reshape(shape = var_189, x = attn_weights_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; + tensor input_7_cast_fp16 = softmax(axis = var_5, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; + tensor attn_output_1_transpose_x_0 = const()[name = tensor("attn_output_1_transpose_x_0"), val = tensor(false)]; + tensor attn_output_1_transpose_y_0 = const()[name = tensor("attn_output_1_transpose_y_0"), val = tensor(false)]; + tensor attn_output_1_cast_fp16 = matmul(transpose_x = attn_output_1_transpose_x_0, transpose_y = attn_output_1_transpose_y_0, x = input_7_cast_fp16, y = value_states_3_cast_fp16)[name = tensor("attn_output_1_cast_fp16")]; + tensor var_194 = const()[name = tensor("op_194"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_3_cast_fp16 = reshape(shape = var_194, x = attn_output_1_cast_fp16)[name = tensor("attn_output_3_cast_fp16")]; + tensor attn_output_5_perm_0 = const()[name = tensor("attn_output_5_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_197 = const()[name = tensor("op_197"), val = tensor([1, 77, 1280])]; + tensor attn_output_5_cast_fp16 = transpose(perm = attn_output_5_perm_0, x = attn_output_3_cast_fp16)[name = tensor("transpose_125")]; + tensor input_9_cast_fp16 = reshape(shape = var_197, x = attn_output_5_cast_fp16)[name = tensor("input_9_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136537472)))]; + tensor text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139814336)))]; + tensor linear_3_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("linear_3_cast_fp16")]; + tensor input_11_cast_fp16 = add(x = input_3_cast_fp16, y = linear_3_cast_fp16)[name = tensor("input_11_cast_fp16")]; + tensor input_13_axes_0 = const()[name = tensor("input_13_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139816960)))]; + tensor text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139819584)))]; + tensor input_13_cast_fp16 = layer_norm(axes = input_13_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(139822208)))]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152929472)))]; + tensor linear_4_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("linear_4_cast_fp16")]; + tensor input_17_mode_0 = const()[name = tensor("input_17_mode_0"), val = tensor("EXACT")]; + tensor input_17_cast_fp16 = gelu(mode = input_17_mode_0, x = linear_4_cast_fp16)[name = tensor("input_17_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152939776)))]; + tensor text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166047040)))]; + tensor linear_5_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("linear_5_cast_fp16")]; + tensor input_19_cast_fp16 = add(x = input_11_cast_fp16, y = linear_5_cast_fp16)[name = tensor("input_19_cast_fp16")]; + tensor hidden_states_7_axes_0 = const()[name = tensor("hidden_states_7_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166049664)))]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166052288)))]; + tensor hidden_states_7_cast_fp16 = layer_norm(axes = hidden_states_7_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("hidden_states_7_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166054912)))]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169331776)))]; + tensor linear_6_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor("linear_6_cast_fp16")]; + tensor var_236_to_fp16 = const()[name = tensor("op_236_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_11_cast_fp16 = mul(x = linear_6_cast_fp16, y = var_236_to_fp16)[name = tensor("tensor_11_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169334400)))]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172611264)))]; + tensor linear_7_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor("linear_7_cast_fp16")]; + tensor var_241 = const()[name = tensor("op_241"), val = tensor([1, -1, 20, 64])]; + tensor var_242_cast_fp16 = reshape(shape = var_241, x = linear_7_cast_fp16)[name = tensor("op_242_cast_fp16")]; + tensor var_243_perm_0 = const()[name = tensor("op_243_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172613888)))]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175890752)))]; + tensor linear_8_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor("linear_8_cast_fp16")]; + tensor var_248 = const()[name = tensor("op_248"), val = tensor([1, -1, 20, 64])]; + tensor var_249_cast_fp16 = reshape(shape = var_248, x = linear_8_cast_fp16)[name = tensor("op_249_cast_fp16")]; + tensor var_250_perm_0 = const()[name = tensor("op_250_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_257 = const()[name = tensor("op_257"), val = tensor([1, 77, 20, 64])]; + tensor var_258_cast_fp16 = reshape(shape = var_257, x = tensor_11_cast_fp16)[name = tensor("op_258_cast_fp16")]; + tensor var_259_perm_0 = const()[name = tensor("op_259_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_261 = const()[name = tensor("op_261"), val = tensor([20, -1, 64])]; + tensor var_259_cast_fp16 = transpose(perm = var_259_perm_0, x = var_258_cast_fp16)[name = tensor("transpose_122")]; + tensor query_states_3_cast_fp16 = reshape(shape = var_261, x = var_259_cast_fp16)[name = tensor("query_states_3_cast_fp16")]; + tensor var_263 = const()[name = tensor("op_263"), val = tensor([20, -1, 64])]; + tensor var_243_cast_fp16 = transpose(perm = var_243_perm_0, x = var_242_cast_fp16)[name = tensor("transpose_124")]; + tensor key_states_7_cast_fp16 = reshape(shape = var_263, x = var_243_cast_fp16)[name = tensor("key_states_7_cast_fp16")]; + tensor var_265 = const()[name = tensor("op_265"), val = tensor([20, -1, 64])]; + tensor var_250_cast_fp16 = transpose(perm = var_250_perm_0, x = var_249_cast_fp16)[name = tensor("transpose_123")]; + tensor value_states_7_cast_fp16 = reshape(shape = var_265, x = var_250_cast_fp16)[name = tensor("value_states_7_cast_fp16")]; + tensor attn_weights_7_transpose_x_1 = const()[name = tensor("attn_weights_7_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_7_transpose_y_1 = const()[name = tensor("attn_weights_7_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_7_cast_fp16 = matmul(transpose_x = attn_weights_7_transpose_x_1, transpose_y = attn_weights_7_transpose_y_1, x = query_states_3_cast_fp16, y = key_states_7_cast_fp16)[name = tensor("attn_weights_7_cast_fp16")]; + tensor var_270 = const()[name = tensor("op_270"), val = tensor([1, 20, 77, 77])]; + tensor var_271_cast_fp16 = reshape(shape = var_270, x = attn_weights_7_cast_fp16)[name = tensor("op_271_cast_fp16")]; + tensor attn_weights_9_cast_fp16 = add(x = var_271_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_9_cast_fp16")]; + tensor var_276 = const()[name = tensor("op_276"), val = tensor([20, 77, 77])]; + tensor input_21_cast_fp16 = reshape(shape = var_276, x = attn_weights_9_cast_fp16)[name = tensor("input_21_cast_fp16")]; + tensor input_23_cast_fp16 = softmax(axis = var_5, x = input_21_cast_fp16)[name = tensor("input_23_cast_fp16")]; + tensor attn_output_7_transpose_x_0 = const()[name = tensor("attn_output_7_transpose_x_0"), val = tensor(false)]; + tensor attn_output_7_transpose_y_0 = const()[name = tensor("attn_output_7_transpose_y_0"), val = tensor(false)]; + tensor attn_output_7_cast_fp16 = matmul(transpose_x = attn_output_7_transpose_x_0, transpose_y = attn_output_7_transpose_y_0, x = input_23_cast_fp16, y = value_states_7_cast_fp16)[name = tensor("attn_output_7_cast_fp16")]; + tensor var_281 = const()[name = tensor("op_281"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_9_cast_fp16 = reshape(shape = var_281, x = attn_output_7_cast_fp16)[name = tensor("attn_output_9_cast_fp16")]; + tensor attn_output_11_perm_0 = const()[name = tensor("attn_output_11_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_284 = const()[name = tensor("op_284"), val = tensor([1, 77, 1280])]; + tensor attn_output_11_cast_fp16 = transpose(perm = attn_output_11_perm_0, x = attn_output_9_cast_fp16)[name = tensor("transpose_121")]; + tensor input_25_cast_fp16 = reshape(shape = var_284, x = attn_output_11_cast_fp16)[name = tensor("input_25_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175893376)))]; + tensor text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179170240)))]; + tensor linear_9_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_9_cast_fp16")]; + tensor input_27_cast_fp16 = add(x = input_19_cast_fp16, y = linear_9_cast_fp16)[name = tensor("input_27_cast_fp16")]; + tensor input_29_axes_0 = const()[name = tensor("input_29_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179172864)))]; + tensor text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179175488)))]; + tensor input_29_cast_fp16 = layer_norm(axes = input_29_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16, x = input_27_cast_fp16)[name = tensor("input_29_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179178112)))]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192285376)))]; + tensor linear_10_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16, x = input_29_cast_fp16)[name = tensor("linear_10_cast_fp16")]; + tensor input_33_mode_0 = const()[name = tensor("input_33_mode_0"), val = tensor("EXACT")]; + tensor input_33_cast_fp16 = gelu(mode = input_33_mode_0, x = linear_10_cast_fp16)[name = tensor("input_33_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192295680)))]; + tensor text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205402944)))]; + tensor linear_11_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16, x = input_33_cast_fp16)[name = tensor("linear_11_cast_fp16")]; + tensor input_35_cast_fp16 = add(x = input_27_cast_fp16, y = linear_11_cast_fp16)[name = tensor("input_35_cast_fp16")]; + tensor hidden_states_13_axes_0 = const()[name = tensor("hidden_states_13_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205405568)))]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205408192)))]; + tensor hidden_states_13_cast_fp16 = layer_norm(axes = hidden_states_13_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("hidden_states_13_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205410816)))]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208687680)))]; + tensor linear_12_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor("linear_12_cast_fp16")]; + tensor var_323_to_fp16 = const()[name = tensor("op_323_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_17_cast_fp16 = mul(x = linear_12_cast_fp16, y = var_323_to_fp16)[name = tensor("tensor_17_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208690304)))]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211967168)))]; + tensor linear_13_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor("linear_13_cast_fp16")]; + tensor var_328 = const()[name = tensor("op_328"), val = tensor([1, -1, 20, 64])]; + tensor var_329_cast_fp16 = reshape(shape = var_328, x = linear_13_cast_fp16)[name = tensor("op_329_cast_fp16")]; + tensor var_330_perm_0 = const()[name = tensor("op_330_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211969792)))]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215246656)))]; + tensor linear_14_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor("linear_14_cast_fp16")]; + tensor var_335 = const()[name = tensor("op_335"), val = tensor([1, -1, 20, 64])]; + tensor var_336_cast_fp16 = reshape(shape = var_335, x = linear_14_cast_fp16)[name = tensor("op_336_cast_fp16")]; + tensor var_337_perm_0 = const()[name = tensor("op_337_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_344 = const()[name = tensor("op_344"), val = tensor([1, 77, 20, 64])]; + tensor var_345_cast_fp16 = reshape(shape = var_344, x = tensor_17_cast_fp16)[name = tensor("op_345_cast_fp16")]; + tensor var_346_perm_0 = const()[name = tensor("op_346_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_348 = const()[name = tensor("op_348"), val = tensor([20, -1, 64])]; + tensor var_346_cast_fp16 = transpose(perm = var_346_perm_0, x = var_345_cast_fp16)[name = tensor("transpose_118")]; + tensor query_states_5_cast_fp16 = reshape(shape = var_348, x = var_346_cast_fp16)[name = tensor("query_states_5_cast_fp16")]; + tensor var_350 = const()[name = tensor("op_350"), val = tensor([20, -1, 64])]; + tensor var_330_cast_fp16 = transpose(perm = var_330_perm_0, x = var_329_cast_fp16)[name = tensor("transpose_120")]; + tensor key_states_11_cast_fp16 = reshape(shape = var_350, x = var_330_cast_fp16)[name = tensor("key_states_11_cast_fp16")]; + tensor var_352 = const()[name = tensor("op_352"), val = tensor([20, -1, 64])]; + tensor var_337_cast_fp16 = transpose(perm = var_337_perm_0, x = var_336_cast_fp16)[name = tensor("transpose_119")]; + tensor value_states_11_cast_fp16 = reshape(shape = var_352, x = var_337_cast_fp16)[name = tensor("value_states_11_cast_fp16")]; + tensor attn_weights_13_transpose_x_1 = const()[name = tensor("attn_weights_13_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_13_transpose_y_1 = const()[name = tensor("attn_weights_13_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_13_cast_fp16 = matmul(transpose_x = attn_weights_13_transpose_x_1, transpose_y = attn_weights_13_transpose_y_1, x = query_states_5_cast_fp16, y = key_states_11_cast_fp16)[name = tensor("attn_weights_13_cast_fp16")]; + tensor var_357 = const()[name = tensor("op_357"), val = tensor([1, 20, 77, 77])]; + tensor var_358_cast_fp16 = reshape(shape = var_357, x = attn_weights_13_cast_fp16)[name = tensor("op_358_cast_fp16")]; + tensor attn_weights_15_cast_fp16 = add(x = var_358_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_15_cast_fp16")]; + tensor var_363 = const()[name = tensor("op_363"), val = tensor([20, 77, 77])]; + tensor input_37_cast_fp16 = reshape(shape = var_363, x = attn_weights_15_cast_fp16)[name = tensor("input_37_cast_fp16")]; + tensor input_39_cast_fp16 = softmax(axis = var_5, x = input_37_cast_fp16)[name = tensor("input_39_cast_fp16")]; + tensor attn_output_13_transpose_x_0 = const()[name = tensor("attn_output_13_transpose_x_0"), val = tensor(false)]; + tensor attn_output_13_transpose_y_0 = const()[name = tensor("attn_output_13_transpose_y_0"), val = tensor(false)]; + tensor attn_output_13_cast_fp16 = matmul(transpose_x = attn_output_13_transpose_x_0, transpose_y = attn_output_13_transpose_y_0, x = input_39_cast_fp16, y = value_states_11_cast_fp16)[name = tensor("attn_output_13_cast_fp16")]; + tensor var_368 = const()[name = tensor("op_368"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_15_cast_fp16 = reshape(shape = var_368, x = attn_output_13_cast_fp16)[name = tensor("attn_output_15_cast_fp16")]; + tensor attn_output_17_perm_0 = const()[name = tensor("attn_output_17_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_371 = const()[name = tensor("op_371"), val = tensor([1, 77, 1280])]; + tensor attn_output_17_cast_fp16 = transpose(perm = attn_output_17_perm_0, x = attn_output_15_cast_fp16)[name = tensor("transpose_117")]; + tensor input_41_cast_fp16 = reshape(shape = var_371, x = attn_output_17_cast_fp16)[name = tensor("input_41_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215249280)))]; + tensor text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218526144)))]; + tensor linear_15_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("linear_15_cast_fp16")]; + tensor input_43_cast_fp16 = add(x = input_35_cast_fp16, y = linear_15_cast_fp16)[name = tensor("input_43_cast_fp16")]; + tensor input_45_axes_0 = const()[name = tensor("input_45_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218528768)))]; + tensor text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218531392)))]; + tensor input_45_cast_fp16 = layer_norm(axes = input_45_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("input_45_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(218534016)))]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231641280)))]; + tensor linear_16_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16, x = input_45_cast_fp16)[name = tensor("linear_16_cast_fp16")]; + tensor input_49_mode_0 = const()[name = tensor("input_49_mode_0"), val = tensor("EXACT")]; + tensor input_49_cast_fp16 = gelu(mode = input_49_mode_0, x = linear_16_cast_fp16)[name = tensor("input_49_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(231651584)))]; + tensor text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244758848)))]; + tensor linear_17_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16, x = input_49_cast_fp16)[name = tensor("linear_17_cast_fp16")]; + tensor input_51_cast_fp16 = add(x = input_43_cast_fp16, y = linear_17_cast_fp16)[name = tensor("input_51_cast_fp16")]; + tensor hidden_states_19_axes_0 = const()[name = tensor("hidden_states_19_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244761472)))]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244764096)))]; + tensor hidden_states_19_cast_fp16 = layer_norm(axes = hidden_states_19_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16, x = input_51_cast_fp16)[name = tensor("hidden_states_19_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(244766720)))]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248043584)))]; + tensor linear_18_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor("linear_18_cast_fp16")]; + tensor var_410_to_fp16 = const()[name = tensor("op_410_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_23_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_410_to_fp16)[name = tensor("tensor_23_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(248046208)))]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251323072)))]; + tensor linear_19_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor("linear_19_cast_fp16")]; + tensor var_415 = const()[name = tensor("op_415"), val = tensor([1, -1, 20, 64])]; + tensor var_416_cast_fp16 = reshape(shape = var_415, x = linear_19_cast_fp16)[name = tensor("op_416_cast_fp16")]; + tensor var_417_perm_0 = const()[name = tensor("op_417_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(251325696)))]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254602560)))]; + tensor linear_20_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor("linear_20_cast_fp16")]; + tensor var_422 = const()[name = tensor("op_422"), val = tensor([1, -1, 20, 64])]; + tensor var_423_cast_fp16 = reshape(shape = var_422, x = linear_20_cast_fp16)[name = tensor("op_423_cast_fp16")]; + tensor var_424_perm_0 = const()[name = tensor("op_424_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_431 = const()[name = tensor("op_431"), val = tensor([1, 77, 20, 64])]; + tensor var_432_cast_fp16 = reshape(shape = var_431, x = tensor_23_cast_fp16)[name = tensor("op_432_cast_fp16")]; + tensor var_433_perm_0 = const()[name = tensor("op_433_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_435 = const()[name = tensor("op_435"), val = tensor([20, -1, 64])]; + tensor var_433_cast_fp16 = transpose(perm = var_433_perm_0, x = var_432_cast_fp16)[name = tensor("transpose_114")]; + tensor query_states_7_cast_fp16 = reshape(shape = var_435, x = var_433_cast_fp16)[name = tensor("query_states_7_cast_fp16")]; + tensor var_437 = const()[name = tensor("op_437"), val = tensor([20, -1, 64])]; + tensor var_417_cast_fp16 = transpose(perm = var_417_perm_0, x = var_416_cast_fp16)[name = tensor("transpose_116")]; + tensor key_states_15_cast_fp16 = reshape(shape = var_437, x = var_417_cast_fp16)[name = tensor("key_states_15_cast_fp16")]; + tensor var_439 = const()[name = tensor("op_439"), val = tensor([20, -1, 64])]; + tensor var_424_cast_fp16 = transpose(perm = var_424_perm_0, x = var_423_cast_fp16)[name = tensor("transpose_115")]; + tensor value_states_15_cast_fp16 = reshape(shape = var_439, x = var_424_cast_fp16)[name = tensor("value_states_15_cast_fp16")]; + tensor attn_weights_19_transpose_x_1 = const()[name = tensor("attn_weights_19_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_19_transpose_y_1 = const()[name = tensor("attn_weights_19_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_19_cast_fp16 = matmul(transpose_x = attn_weights_19_transpose_x_1, transpose_y = attn_weights_19_transpose_y_1, x = query_states_7_cast_fp16, y = key_states_15_cast_fp16)[name = tensor("attn_weights_19_cast_fp16")]; + tensor var_444 = const()[name = tensor("op_444"), val = tensor([1, 20, 77, 77])]; + tensor var_445_cast_fp16 = reshape(shape = var_444, x = attn_weights_19_cast_fp16)[name = tensor("op_445_cast_fp16")]; + tensor attn_weights_21_cast_fp16 = add(x = var_445_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_21_cast_fp16")]; + tensor var_450 = const()[name = tensor("op_450"), val = tensor([20, 77, 77])]; + tensor input_53_cast_fp16 = reshape(shape = var_450, x = attn_weights_21_cast_fp16)[name = tensor("input_53_cast_fp16")]; + tensor input_55_cast_fp16 = softmax(axis = var_5, x = input_53_cast_fp16)[name = tensor("input_55_cast_fp16")]; + tensor attn_output_19_transpose_x_0 = const()[name = tensor("attn_output_19_transpose_x_0"), val = tensor(false)]; + tensor attn_output_19_transpose_y_0 = const()[name = tensor("attn_output_19_transpose_y_0"), val = tensor(false)]; + tensor attn_output_19_cast_fp16 = matmul(transpose_x = attn_output_19_transpose_x_0, transpose_y = attn_output_19_transpose_y_0, x = input_55_cast_fp16, y = value_states_15_cast_fp16)[name = tensor("attn_output_19_cast_fp16")]; + tensor var_455 = const()[name = tensor("op_455"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_21_cast_fp16 = reshape(shape = var_455, x = attn_output_19_cast_fp16)[name = tensor("attn_output_21_cast_fp16")]; + tensor attn_output_23_perm_0 = const()[name = tensor("attn_output_23_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_458 = const()[name = tensor("op_458"), val = tensor([1, 77, 1280])]; + tensor attn_output_23_cast_fp16 = transpose(perm = attn_output_23_perm_0, x = attn_output_21_cast_fp16)[name = tensor("transpose_113")]; + tensor input_57_cast_fp16 = reshape(shape = var_458, x = attn_output_23_cast_fp16)[name = tensor("input_57_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(254605184)))]; + tensor text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257882048)))]; + tensor linear_21_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("linear_21_cast_fp16")]; + tensor input_59_cast_fp16 = add(x = input_51_cast_fp16, y = linear_21_cast_fp16)[name = tensor("input_59_cast_fp16")]; + tensor input_61_axes_0 = const()[name = tensor("input_61_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257884672)))]; + tensor text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257887296)))]; + tensor input_61_cast_fp16 = layer_norm(axes = input_61_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("input_61_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(257889920)))]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(270997184)))]; + tensor linear_22_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16, x = input_61_cast_fp16)[name = tensor("linear_22_cast_fp16")]; + tensor input_65_mode_0 = const()[name = tensor("input_65_mode_0"), val = tensor("EXACT")]; + tensor input_65_cast_fp16 = gelu(mode = input_65_mode_0, x = linear_22_cast_fp16)[name = tensor("input_65_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(271007488)))]; + tensor text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284114752)))]; + tensor linear_23_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16, x = input_65_cast_fp16)[name = tensor("linear_23_cast_fp16")]; + tensor input_67_cast_fp16 = add(x = input_59_cast_fp16, y = linear_23_cast_fp16)[name = tensor("input_67_cast_fp16")]; + tensor hidden_states_25_axes_0 = const()[name = tensor("hidden_states_25_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284117376)))]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284120000)))]; + tensor hidden_states_25_cast_fp16 = layer_norm(axes = hidden_states_25_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16, x = input_67_cast_fp16)[name = tensor("hidden_states_25_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(284122624)))]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287399488)))]; + tensor linear_24_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor("linear_24_cast_fp16")]; + tensor var_497_to_fp16 = const()[name = tensor("op_497_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_29_cast_fp16 = mul(x = linear_24_cast_fp16, y = var_497_to_fp16)[name = tensor("tensor_29_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(287402112)))]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290678976)))]; + tensor linear_25_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor("linear_25_cast_fp16")]; + tensor var_502 = const()[name = tensor("op_502"), val = tensor([1, -1, 20, 64])]; + tensor var_503_cast_fp16 = reshape(shape = var_502, x = linear_25_cast_fp16)[name = tensor("op_503_cast_fp16")]; + tensor var_504_perm_0 = const()[name = tensor("op_504_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(290681600)))]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(293958464)))]; + tensor linear_26_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor("linear_26_cast_fp16")]; + tensor var_509 = const()[name = tensor("op_509"), val = tensor([1, -1, 20, 64])]; + tensor var_510_cast_fp16 = reshape(shape = var_509, x = linear_26_cast_fp16)[name = tensor("op_510_cast_fp16")]; + tensor var_511_perm_0 = const()[name = tensor("op_511_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_518 = const()[name = tensor("op_518"), val = tensor([1, 77, 20, 64])]; + tensor var_519_cast_fp16 = reshape(shape = var_518, x = tensor_29_cast_fp16)[name = tensor("op_519_cast_fp16")]; + tensor var_520_perm_0 = const()[name = tensor("op_520_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_522 = const()[name = tensor("op_522"), val = tensor([20, -1, 64])]; + tensor var_520_cast_fp16 = transpose(perm = var_520_perm_0, x = var_519_cast_fp16)[name = tensor("transpose_110")]; + tensor query_states_9_cast_fp16 = reshape(shape = var_522, x = var_520_cast_fp16)[name = tensor("query_states_9_cast_fp16")]; + tensor var_524 = const()[name = tensor("op_524"), val = tensor([20, -1, 64])]; + tensor var_504_cast_fp16 = transpose(perm = var_504_perm_0, x = var_503_cast_fp16)[name = tensor("transpose_112")]; + tensor key_states_19_cast_fp16 = reshape(shape = var_524, x = var_504_cast_fp16)[name = tensor("key_states_19_cast_fp16")]; + tensor var_526 = const()[name = tensor("op_526"), val = tensor([20, -1, 64])]; + tensor var_511_cast_fp16 = transpose(perm = var_511_perm_0, x = var_510_cast_fp16)[name = tensor("transpose_111")]; + tensor value_states_19_cast_fp16 = reshape(shape = var_526, x = var_511_cast_fp16)[name = tensor("value_states_19_cast_fp16")]; + tensor attn_weights_25_transpose_x_1 = const()[name = tensor("attn_weights_25_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_25_transpose_y_1 = const()[name = tensor("attn_weights_25_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_25_cast_fp16 = matmul(transpose_x = attn_weights_25_transpose_x_1, transpose_y = attn_weights_25_transpose_y_1, x = query_states_9_cast_fp16, y = key_states_19_cast_fp16)[name = tensor("attn_weights_25_cast_fp16")]; + tensor var_531 = const()[name = tensor("op_531"), val = tensor([1, 20, 77, 77])]; + tensor var_532_cast_fp16 = reshape(shape = var_531, x = attn_weights_25_cast_fp16)[name = tensor("op_532_cast_fp16")]; + tensor attn_weights_27_cast_fp16 = add(x = var_532_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_27_cast_fp16")]; + tensor var_537 = const()[name = tensor("op_537"), val = tensor([20, 77, 77])]; + tensor input_69_cast_fp16 = reshape(shape = var_537, x = attn_weights_27_cast_fp16)[name = tensor("input_69_cast_fp16")]; + tensor input_71_cast_fp16 = softmax(axis = var_5, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; + tensor attn_output_25_transpose_x_0 = const()[name = tensor("attn_output_25_transpose_x_0"), val = tensor(false)]; + tensor attn_output_25_transpose_y_0 = const()[name = tensor("attn_output_25_transpose_y_0"), val = tensor(false)]; + tensor attn_output_25_cast_fp16 = matmul(transpose_x = attn_output_25_transpose_x_0, transpose_y = attn_output_25_transpose_y_0, x = input_71_cast_fp16, y = value_states_19_cast_fp16)[name = tensor("attn_output_25_cast_fp16")]; + tensor var_542 = const()[name = tensor("op_542"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_27_cast_fp16 = reshape(shape = var_542, x = attn_output_25_cast_fp16)[name = tensor("attn_output_27_cast_fp16")]; + tensor attn_output_29_perm_0 = const()[name = tensor("attn_output_29_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_545 = const()[name = tensor("op_545"), val = tensor([1, 77, 1280])]; + tensor attn_output_29_cast_fp16 = transpose(perm = attn_output_29_perm_0, x = attn_output_27_cast_fp16)[name = tensor("transpose_109")]; + tensor input_73_cast_fp16 = reshape(shape = var_545, x = attn_output_29_cast_fp16)[name = tensor("input_73_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(293961088)))]; + tensor text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297237952)))]; + tensor linear_27_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16, x = input_73_cast_fp16)[name = tensor("linear_27_cast_fp16")]; + tensor input_75_cast_fp16 = add(x = input_67_cast_fp16, y = linear_27_cast_fp16)[name = tensor("input_75_cast_fp16")]; + tensor input_77_axes_0 = const()[name = tensor("input_77_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297240576)))]; + tensor text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297243200)))]; + tensor input_77_cast_fp16 = layer_norm(axes = input_77_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16, x = input_75_cast_fp16)[name = tensor("input_77_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(297245824)))]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310353088)))]; + tensor linear_28_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("linear_28_cast_fp16")]; + tensor input_81_mode_0 = const()[name = tensor("input_81_mode_0"), val = tensor("EXACT")]; + tensor input_81_cast_fp16 = gelu(mode = input_81_mode_0, x = linear_28_cast_fp16)[name = tensor("input_81_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(310363392)))]; + tensor text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323470656)))]; + tensor linear_29_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16, x = input_81_cast_fp16)[name = tensor("linear_29_cast_fp16")]; + tensor input_83_cast_fp16 = add(x = input_75_cast_fp16, y = linear_29_cast_fp16)[name = tensor("input_83_cast_fp16")]; + tensor hidden_states_31_axes_0 = const()[name = tensor("hidden_states_31_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323473280)))]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323475904)))]; + tensor hidden_states_31_cast_fp16 = layer_norm(axes = hidden_states_31_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("hidden_states_31_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(323478528)))]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326755392)))]; + tensor linear_30_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor("linear_30_cast_fp16")]; + tensor var_584_to_fp16 = const()[name = tensor("op_584_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_35_cast_fp16 = mul(x = linear_30_cast_fp16, y = var_584_to_fp16)[name = tensor("tensor_35_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(326758016)))]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330034880)))]; + tensor linear_31_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor("linear_31_cast_fp16")]; + tensor var_589 = const()[name = tensor("op_589"), val = tensor([1, -1, 20, 64])]; + tensor var_590_cast_fp16 = reshape(shape = var_589, x = linear_31_cast_fp16)[name = tensor("op_590_cast_fp16")]; + tensor var_591_perm_0 = const()[name = tensor("op_591_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(330037504)))]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333314368)))]; + tensor linear_32_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor("linear_32_cast_fp16")]; + tensor var_596 = const()[name = tensor("op_596"), val = tensor([1, -1, 20, 64])]; + tensor var_597_cast_fp16 = reshape(shape = var_596, x = linear_32_cast_fp16)[name = tensor("op_597_cast_fp16")]; + tensor var_598_perm_0 = const()[name = tensor("op_598_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_605 = const()[name = tensor("op_605"), val = tensor([1, 77, 20, 64])]; + tensor var_606_cast_fp16 = reshape(shape = var_605, x = tensor_35_cast_fp16)[name = tensor("op_606_cast_fp16")]; + tensor var_607_perm_0 = const()[name = tensor("op_607_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_609 = const()[name = tensor("op_609"), val = tensor([20, -1, 64])]; + tensor var_607_cast_fp16 = transpose(perm = var_607_perm_0, x = var_606_cast_fp16)[name = tensor("transpose_106")]; + tensor query_states_11_cast_fp16 = reshape(shape = var_609, x = var_607_cast_fp16)[name = tensor("query_states_11_cast_fp16")]; + tensor var_611 = const()[name = tensor("op_611"), val = tensor([20, -1, 64])]; + tensor var_591_cast_fp16 = transpose(perm = var_591_perm_0, x = var_590_cast_fp16)[name = tensor("transpose_108")]; + tensor key_states_23_cast_fp16 = reshape(shape = var_611, x = var_591_cast_fp16)[name = tensor("key_states_23_cast_fp16")]; + tensor var_613 = const()[name = tensor("op_613"), val = tensor([20, -1, 64])]; + tensor var_598_cast_fp16 = transpose(perm = var_598_perm_0, x = var_597_cast_fp16)[name = tensor("transpose_107")]; + tensor value_states_23_cast_fp16 = reshape(shape = var_613, x = var_598_cast_fp16)[name = tensor("value_states_23_cast_fp16")]; + tensor attn_weights_31_transpose_x_1 = const()[name = tensor("attn_weights_31_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_31_transpose_y_1 = const()[name = tensor("attn_weights_31_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_31_cast_fp16 = matmul(transpose_x = attn_weights_31_transpose_x_1, transpose_y = attn_weights_31_transpose_y_1, x = query_states_11_cast_fp16, y = key_states_23_cast_fp16)[name = tensor("attn_weights_31_cast_fp16")]; + tensor var_618 = const()[name = tensor("op_618"), val = tensor([1, 20, 77, 77])]; + tensor var_619_cast_fp16 = reshape(shape = var_618, x = attn_weights_31_cast_fp16)[name = tensor("op_619_cast_fp16")]; + tensor attn_weights_33_cast_fp16 = add(x = var_619_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_33_cast_fp16")]; + tensor var_624 = const()[name = tensor("op_624"), val = tensor([20, 77, 77])]; + tensor input_85_cast_fp16 = reshape(shape = var_624, x = attn_weights_33_cast_fp16)[name = tensor("input_85_cast_fp16")]; + tensor input_87_cast_fp16 = softmax(axis = var_5, x = input_85_cast_fp16)[name = tensor("input_87_cast_fp16")]; + tensor attn_output_31_transpose_x_0 = const()[name = tensor("attn_output_31_transpose_x_0"), val = tensor(false)]; + tensor attn_output_31_transpose_y_0 = const()[name = tensor("attn_output_31_transpose_y_0"), val = tensor(false)]; + tensor attn_output_31_cast_fp16 = matmul(transpose_x = attn_output_31_transpose_x_0, transpose_y = attn_output_31_transpose_y_0, x = input_87_cast_fp16, y = value_states_23_cast_fp16)[name = tensor("attn_output_31_cast_fp16")]; + tensor var_629 = const()[name = tensor("op_629"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_33_cast_fp16 = reshape(shape = var_629, x = attn_output_31_cast_fp16)[name = tensor("attn_output_33_cast_fp16")]; + tensor attn_output_35_perm_0 = const()[name = tensor("attn_output_35_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_632 = const()[name = tensor("op_632"), val = tensor([1, 77, 1280])]; + tensor attn_output_35_cast_fp16 = transpose(perm = attn_output_35_perm_0, x = attn_output_33_cast_fp16)[name = tensor("transpose_105")]; + tensor input_89_cast_fp16 = reshape(shape = var_632, x = attn_output_35_cast_fp16)[name = tensor("input_89_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(333316992)))]; + tensor text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336593856)))]; + tensor linear_33_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16, x = input_89_cast_fp16)[name = tensor("linear_33_cast_fp16")]; + tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_33_cast_fp16)[name = tensor("input_91_cast_fp16")]; + tensor input_93_axes_0 = const()[name = tensor("input_93_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336596480)))]; + tensor text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336599104)))]; + tensor input_93_cast_fp16 = layer_norm(axes = input_93_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("input_93_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(336601728)))]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349708992)))]; + tensor linear_34_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("linear_34_cast_fp16")]; + tensor input_97_mode_0 = const()[name = tensor("input_97_mode_0"), val = tensor("EXACT")]; + tensor input_97_cast_fp16 = gelu(mode = input_97_mode_0, x = linear_34_cast_fp16)[name = tensor("input_97_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(349719296)))]; + tensor text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362826560)))]; + tensor linear_35_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16, x = input_97_cast_fp16)[name = tensor("linear_35_cast_fp16")]; + tensor input_99_cast_fp16 = add(x = input_91_cast_fp16, y = linear_35_cast_fp16)[name = tensor("input_99_cast_fp16")]; + tensor hidden_states_37_axes_0 = const()[name = tensor("hidden_states_37_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362829184)))]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362831808)))]; + tensor hidden_states_37_cast_fp16 = layer_norm(axes = hidden_states_37_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("hidden_states_37_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(362834432)))]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(366111296)))]; + tensor linear_36_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor("linear_36_cast_fp16")]; + tensor var_671_to_fp16 = const()[name = tensor("op_671_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_41_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_671_to_fp16)[name = tensor("tensor_41_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(366113920)))]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(369390784)))]; + tensor linear_37_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor("linear_37_cast_fp16")]; + tensor var_676 = const()[name = tensor("op_676"), val = tensor([1, -1, 20, 64])]; + tensor var_677_cast_fp16 = reshape(shape = var_676, x = linear_37_cast_fp16)[name = tensor("op_677_cast_fp16")]; + tensor var_678_perm_0 = const()[name = tensor("op_678_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(369393408)))]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372670272)))]; + tensor linear_38_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor("linear_38_cast_fp16")]; + tensor var_683 = const()[name = tensor("op_683"), val = tensor([1, -1, 20, 64])]; + tensor var_684_cast_fp16 = reshape(shape = var_683, x = linear_38_cast_fp16)[name = tensor("op_684_cast_fp16")]; + tensor var_685_perm_0 = const()[name = tensor("op_685_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_692 = const()[name = tensor("op_692"), val = tensor([1, 77, 20, 64])]; + tensor var_693_cast_fp16 = reshape(shape = var_692, x = tensor_41_cast_fp16)[name = tensor("op_693_cast_fp16")]; + tensor var_694_perm_0 = const()[name = tensor("op_694_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_696 = const()[name = tensor("op_696"), val = tensor([20, -1, 64])]; + tensor var_694_cast_fp16 = transpose(perm = var_694_perm_0, x = var_693_cast_fp16)[name = tensor("transpose_102")]; + tensor query_states_13_cast_fp16 = reshape(shape = var_696, x = var_694_cast_fp16)[name = tensor("query_states_13_cast_fp16")]; + tensor var_698 = const()[name = tensor("op_698"), val = tensor([20, -1, 64])]; + tensor var_678_cast_fp16 = transpose(perm = var_678_perm_0, x = var_677_cast_fp16)[name = tensor("transpose_104")]; + tensor key_states_27_cast_fp16 = reshape(shape = var_698, x = var_678_cast_fp16)[name = tensor("key_states_27_cast_fp16")]; + tensor var_700 = const()[name = tensor("op_700"), val = tensor([20, -1, 64])]; + tensor var_685_cast_fp16 = transpose(perm = var_685_perm_0, x = var_684_cast_fp16)[name = tensor("transpose_103")]; + tensor value_states_27_cast_fp16 = reshape(shape = var_700, x = var_685_cast_fp16)[name = tensor("value_states_27_cast_fp16")]; + tensor attn_weights_37_transpose_x_1 = const()[name = tensor("attn_weights_37_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_37_transpose_y_1 = const()[name = tensor("attn_weights_37_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_37_cast_fp16 = matmul(transpose_x = attn_weights_37_transpose_x_1, transpose_y = attn_weights_37_transpose_y_1, x = query_states_13_cast_fp16, y = key_states_27_cast_fp16)[name = tensor("attn_weights_37_cast_fp16")]; + tensor var_705 = const()[name = tensor("op_705"), val = tensor([1, 20, 77, 77])]; + tensor var_706_cast_fp16 = reshape(shape = var_705, x = attn_weights_37_cast_fp16)[name = tensor("op_706_cast_fp16")]; + tensor attn_weights_39_cast_fp16 = add(x = var_706_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_39_cast_fp16")]; + tensor var_711 = const()[name = tensor("op_711"), val = tensor([20, 77, 77])]; + tensor input_101_cast_fp16 = reshape(shape = var_711, x = attn_weights_39_cast_fp16)[name = tensor("input_101_cast_fp16")]; + tensor input_103_cast_fp16 = softmax(axis = var_5, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; + tensor attn_output_37_transpose_x_0 = const()[name = tensor("attn_output_37_transpose_x_0"), val = tensor(false)]; + tensor attn_output_37_transpose_y_0 = const()[name = tensor("attn_output_37_transpose_y_0"), val = tensor(false)]; + tensor attn_output_37_cast_fp16 = matmul(transpose_x = attn_output_37_transpose_x_0, transpose_y = attn_output_37_transpose_y_0, x = input_103_cast_fp16, y = value_states_27_cast_fp16)[name = tensor("attn_output_37_cast_fp16")]; + tensor var_716 = const()[name = tensor("op_716"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_39_cast_fp16 = reshape(shape = var_716, x = attn_output_37_cast_fp16)[name = tensor("attn_output_39_cast_fp16")]; + tensor attn_output_41_perm_0 = const()[name = tensor("attn_output_41_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_719 = const()[name = tensor("op_719"), val = tensor([1, 77, 1280])]; + tensor attn_output_41_cast_fp16 = transpose(perm = attn_output_41_perm_0, x = attn_output_39_cast_fp16)[name = tensor("transpose_101")]; + tensor input_105_cast_fp16 = reshape(shape = var_719, x = attn_output_41_cast_fp16)[name = tensor("input_105_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(372672896)))]; + tensor text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(375949760)))]; + tensor linear_39_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("linear_39_cast_fp16")]; + tensor input_107_cast_fp16 = add(x = input_99_cast_fp16, y = linear_39_cast_fp16)[name = tensor("input_107_cast_fp16")]; + tensor input_109_axes_0 = const()[name = tensor("input_109_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(375952384)))]; + tensor text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(375955008)))]; + tensor input_109_cast_fp16 = layer_norm(axes = input_109_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16, x = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(375957632)))]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389064896)))]; + tensor linear_40_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("linear_40_cast_fp16")]; + tensor input_113_mode_0 = const()[name = tensor("input_113_mode_0"), val = tensor("EXACT")]; + tensor input_113_cast_fp16 = gelu(mode = input_113_mode_0, x = linear_40_cast_fp16)[name = tensor("input_113_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(389075200)))]; + tensor text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402182464)))]; + tensor linear_41_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16, x = input_113_cast_fp16)[name = tensor("linear_41_cast_fp16")]; + tensor input_115_cast_fp16 = add(x = input_107_cast_fp16, y = linear_41_cast_fp16)[name = tensor("input_115_cast_fp16")]; + tensor hidden_states_43_axes_0 = const()[name = tensor("hidden_states_43_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402185088)))]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402187712)))]; + tensor hidden_states_43_cast_fp16 = layer_norm(axes = hidden_states_43_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("hidden_states_43_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(402190336)))]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405467200)))]; + tensor linear_42_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor("linear_42_cast_fp16")]; + tensor var_758_to_fp16 = const()[name = tensor("op_758_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_47_cast_fp16 = mul(x = linear_42_cast_fp16, y = var_758_to_fp16)[name = tensor("tensor_47_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(405469824)))]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408746688)))]; + tensor linear_43_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor("linear_43_cast_fp16")]; + tensor var_763 = const()[name = tensor("op_763"), val = tensor([1, -1, 20, 64])]; + tensor var_764_cast_fp16 = reshape(shape = var_763, x = linear_43_cast_fp16)[name = tensor("op_764_cast_fp16")]; + tensor var_765_perm_0 = const()[name = tensor("op_765_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(408749312)))]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412026176)))]; + tensor linear_44_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor("linear_44_cast_fp16")]; + tensor var_770 = const()[name = tensor("op_770"), val = tensor([1, -1, 20, 64])]; + tensor var_771_cast_fp16 = reshape(shape = var_770, x = linear_44_cast_fp16)[name = tensor("op_771_cast_fp16")]; + tensor var_772_perm_0 = const()[name = tensor("op_772_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_779 = const()[name = tensor("op_779"), val = tensor([1, 77, 20, 64])]; + tensor var_780_cast_fp16 = reshape(shape = var_779, x = tensor_47_cast_fp16)[name = tensor("op_780_cast_fp16")]; + tensor var_781_perm_0 = const()[name = tensor("op_781_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_783 = const()[name = tensor("op_783"), val = tensor([20, -1, 64])]; + tensor var_781_cast_fp16 = transpose(perm = var_781_perm_0, x = var_780_cast_fp16)[name = tensor("transpose_98")]; + tensor query_states_15_cast_fp16 = reshape(shape = var_783, x = var_781_cast_fp16)[name = tensor("query_states_15_cast_fp16")]; + tensor var_785 = const()[name = tensor("op_785"), val = tensor([20, -1, 64])]; + tensor var_765_cast_fp16 = transpose(perm = var_765_perm_0, x = var_764_cast_fp16)[name = tensor("transpose_100")]; + tensor key_states_31_cast_fp16 = reshape(shape = var_785, x = var_765_cast_fp16)[name = tensor("key_states_31_cast_fp16")]; + tensor var_787 = const()[name = tensor("op_787"), val = tensor([20, -1, 64])]; + tensor var_772_cast_fp16 = transpose(perm = var_772_perm_0, x = var_771_cast_fp16)[name = tensor("transpose_99")]; + tensor value_states_31_cast_fp16 = reshape(shape = var_787, x = var_772_cast_fp16)[name = tensor("value_states_31_cast_fp16")]; + tensor attn_weights_43_transpose_x_1 = const()[name = tensor("attn_weights_43_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_43_transpose_y_1 = const()[name = tensor("attn_weights_43_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_43_cast_fp16 = matmul(transpose_x = attn_weights_43_transpose_x_1, transpose_y = attn_weights_43_transpose_y_1, x = query_states_15_cast_fp16, y = key_states_31_cast_fp16)[name = tensor("attn_weights_43_cast_fp16")]; + tensor var_792 = const()[name = tensor("op_792"), val = tensor([1, 20, 77, 77])]; + tensor var_793_cast_fp16 = reshape(shape = var_792, x = attn_weights_43_cast_fp16)[name = tensor("op_793_cast_fp16")]; + tensor attn_weights_45_cast_fp16 = add(x = var_793_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_45_cast_fp16")]; + tensor var_798 = const()[name = tensor("op_798"), val = tensor([20, 77, 77])]; + tensor input_117_cast_fp16 = reshape(shape = var_798, x = attn_weights_45_cast_fp16)[name = tensor("input_117_cast_fp16")]; + tensor input_119_cast_fp16 = softmax(axis = var_5, x = input_117_cast_fp16)[name = tensor("input_119_cast_fp16")]; + tensor attn_output_43_transpose_x_0 = const()[name = tensor("attn_output_43_transpose_x_0"), val = tensor(false)]; + tensor attn_output_43_transpose_y_0 = const()[name = tensor("attn_output_43_transpose_y_0"), val = tensor(false)]; + tensor attn_output_43_cast_fp16 = matmul(transpose_x = attn_output_43_transpose_x_0, transpose_y = attn_output_43_transpose_y_0, x = input_119_cast_fp16, y = value_states_31_cast_fp16)[name = tensor("attn_output_43_cast_fp16")]; + tensor var_803 = const()[name = tensor("op_803"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_45_cast_fp16 = reshape(shape = var_803, x = attn_output_43_cast_fp16)[name = tensor("attn_output_45_cast_fp16")]; + tensor attn_output_47_perm_0 = const()[name = tensor("attn_output_47_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_806 = const()[name = tensor("op_806"), val = tensor([1, 77, 1280])]; + tensor attn_output_47_cast_fp16 = transpose(perm = attn_output_47_perm_0, x = attn_output_45_cast_fp16)[name = tensor("transpose_97")]; + tensor input_121_cast_fp16 = reshape(shape = var_806, x = attn_output_47_cast_fp16)[name = tensor("input_121_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(412028800)))]; + tensor text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(415305664)))]; + tensor linear_45_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("linear_45_cast_fp16")]; + tensor input_123_cast_fp16 = add(x = input_115_cast_fp16, y = linear_45_cast_fp16)[name = tensor("input_123_cast_fp16")]; + tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(415308288)))]; + tensor text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(415310912)))]; + tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(415313536)))]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428420800)))]; + tensor linear_46_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16, x = input_125_cast_fp16)[name = tensor("linear_46_cast_fp16")]; + tensor input_129_mode_0 = const()[name = tensor("input_129_mode_0"), val = tensor("EXACT")]; + tensor input_129_cast_fp16 = gelu(mode = input_129_mode_0, x = linear_46_cast_fp16)[name = tensor("input_129_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(428431104)))]; + tensor text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441538368)))]; + tensor linear_47_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("linear_47_cast_fp16")]; + tensor input_131_cast_fp16 = add(x = input_123_cast_fp16, y = linear_47_cast_fp16)[name = tensor("input_131_cast_fp16")]; + tensor hidden_states_49_axes_0 = const()[name = tensor("hidden_states_49_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441540992)))]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441543616)))]; + tensor hidden_states_49_cast_fp16 = layer_norm(axes = hidden_states_49_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("hidden_states_49_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(441546240)))]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(444823104)))]; + tensor linear_48_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor("linear_48_cast_fp16")]; + tensor var_845_to_fp16 = const()[name = tensor("op_845_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_53_cast_fp16 = mul(x = linear_48_cast_fp16, y = var_845_to_fp16)[name = tensor("tensor_53_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(444825728)))]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(448102592)))]; + tensor linear_49_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor("linear_49_cast_fp16")]; + tensor var_850 = const()[name = tensor("op_850"), val = tensor([1, -1, 20, 64])]; + tensor var_851_cast_fp16 = reshape(shape = var_850, x = linear_49_cast_fp16)[name = tensor("op_851_cast_fp16")]; + tensor var_852_perm_0 = const()[name = tensor("op_852_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(448105216)))]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451382080)))]; + tensor linear_50_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor("linear_50_cast_fp16")]; + tensor var_857 = const()[name = tensor("op_857"), val = tensor([1, -1, 20, 64])]; + tensor var_858_cast_fp16 = reshape(shape = var_857, x = linear_50_cast_fp16)[name = tensor("op_858_cast_fp16")]; + tensor var_859_perm_0 = const()[name = tensor("op_859_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_866 = const()[name = tensor("op_866"), val = tensor([1, 77, 20, 64])]; + tensor var_867_cast_fp16 = reshape(shape = var_866, x = tensor_53_cast_fp16)[name = tensor("op_867_cast_fp16")]; + tensor var_868_perm_0 = const()[name = tensor("op_868_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_870 = const()[name = tensor("op_870"), val = tensor([20, -1, 64])]; + tensor var_868_cast_fp16 = transpose(perm = var_868_perm_0, x = var_867_cast_fp16)[name = tensor("transpose_94")]; + tensor query_states_17_cast_fp16 = reshape(shape = var_870, x = var_868_cast_fp16)[name = tensor("query_states_17_cast_fp16")]; + tensor var_872 = const()[name = tensor("op_872"), val = tensor([20, -1, 64])]; + tensor var_852_cast_fp16 = transpose(perm = var_852_perm_0, x = var_851_cast_fp16)[name = tensor("transpose_96")]; + tensor key_states_35_cast_fp16 = reshape(shape = var_872, x = var_852_cast_fp16)[name = tensor("key_states_35_cast_fp16")]; + tensor var_874 = const()[name = tensor("op_874"), val = tensor([20, -1, 64])]; + tensor var_859_cast_fp16 = transpose(perm = var_859_perm_0, x = var_858_cast_fp16)[name = tensor("transpose_95")]; + tensor value_states_35_cast_fp16 = reshape(shape = var_874, x = var_859_cast_fp16)[name = tensor("value_states_35_cast_fp16")]; + tensor attn_weights_49_transpose_x_1 = const()[name = tensor("attn_weights_49_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_49_transpose_y_1 = const()[name = tensor("attn_weights_49_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_49_cast_fp16 = matmul(transpose_x = attn_weights_49_transpose_x_1, transpose_y = attn_weights_49_transpose_y_1, x = query_states_17_cast_fp16, y = key_states_35_cast_fp16)[name = tensor("attn_weights_49_cast_fp16")]; + tensor var_879 = const()[name = tensor("op_879"), val = tensor([1, 20, 77, 77])]; + tensor var_880_cast_fp16 = reshape(shape = var_879, x = attn_weights_49_cast_fp16)[name = tensor("op_880_cast_fp16")]; + tensor attn_weights_51_cast_fp16 = add(x = var_880_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_51_cast_fp16")]; + tensor var_885 = const()[name = tensor("op_885"), val = tensor([20, 77, 77])]; + tensor input_133_cast_fp16 = reshape(shape = var_885, x = attn_weights_51_cast_fp16)[name = tensor("input_133_cast_fp16")]; + tensor input_135_cast_fp16 = softmax(axis = var_5, x = input_133_cast_fp16)[name = tensor("input_135_cast_fp16")]; + tensor attn_output_49_transpose_x_0 = const()[name = tensor("attn_output_49_transpose_x_0"), val = tensor(false)]; + tensor attn_output_49_transpose_y_0 = const()[name = tensor("attn_output_49_transpose_y_0"), val = tensor(false)]; + tensor attn_output_49_cast_fp16 = matmul(transpose_x = attn_output_49_transpose_x_0, transpose_y = attn_output_49_transpose_y_0, x = input_135_cast_fp16, y = value_states_35_cast_fp16)[name = tensor("attn_output_49_cast_fp16")]; + tensor var_890 = const()[name = tensor("op_890"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_51_cast_fp16 = reshape(shape = var_890, x = attn_output_49_cast_fp16)[name = tensor("attn_output_51_cast_fp16")]; + tensor attn_output_53_perm_0 = const()[name = tensor("attn_output_53_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_893 = const()[name = tensor("op_893"), val = tensor([1, 77, 1280])]; + tensor attn_output_53_cast_fp16 = transpose(perm = attn_output_53_perm_0, x = attn_output_51_cast_fp16)[name = tensor("transpose_93")]; + tensor input_137_cast_fp16 = reshape(shape = var_893, x = attn_output_53_cast_fp16)[name = tensor("input_137_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(451384704)))]; + tensor text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454661568)))]; + tensor linear_51_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16, x = input_137_cast_fp16)[name = tensor("linear_51_cast_fp16")]; + tensor input_139_cast_fp16 = add(x = input_131_cast_fp16, y = linear_51_cast_fp16)[name = tensor("input_139_cast_fp16")]; + tensor input_141_axes_0 = const()[name = tensor("input_141_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454664192)))]; + tensor text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454666816)))]; + tensor input_141_cast_fp16 = layer_norm(axes = input_141_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("input_141_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(454669440)))]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(467776704)))]; + tensor linear_52_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16, x = input_141_cast_fp16)[name = tensor("linear_52_cast_fp16")]; + tensor input_145_mode_0 = const()[name = tensor("input_145_mode_0"), val = tensor("EXACT")]; + tensor input_145_cast_fp16 = gelu(mode = input_145_mode_0, x = linear_52_cast_fp16)[name = tensor("input_145_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(467787008)))]; + tensor text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480894272)))]; + tensor linear_53_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("linear_53_cast_fp16")]; + tensor input_147_cast_fp16 = add(x = input_139_cast_fp16, y = linear_53_cast_fp16)[name = tensor("input_147_cast_fp16")]; + tensor hidden_states_55_axes_0 = const()[name = tensor("hidden_states_55_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480896896)))]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480899520)))]; + tensor hidden_states_55_cast_fp16 = layer_norm(axes = hidden_states_55_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("hidden_states_55_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(480902144)))]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(484179008)))]; + tensor linear_54_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor("linear_54_cast_fp16")]; + tensor var_932_to_fp16 = const()[name = tensor("op_932_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_59_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_932_to_fp16)[name = tensor("tensor_59_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(484181632)))]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(487458496)))]; + tensor linear_55_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor("linear_55_cast_fp16")]; + tensor var_937 = const()[name = tensor("op_937"), val = tensor([1, -1, 20, 64])]; + tensor var_938_cast_fp16 = reshape(shape = var_937, x = linear_55_cast_fp16)[name = tensor("op_938_cast_fp16")]; + tensor var_939_perm_0 = const()[name = tensor("op_939_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(487461120)))]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490737984)))]; + tensor linear_56_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor("linear_56_cast_fp16")]; + tensor var_944 = const()[name = tensor("op_944"), val = tensor([1, -1, 20, 64])]; + tensor var_945_cast_fp16 = reshape(shape = var_944, x = linear_56_cast_fp16)[name = tensor("op_945_cast_fp16")]; + tensor var_946_perm_0 = const()[name = tensor("op_946_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_953 = const()[name = tensor("op_953"), val = tensor([1, 77, 20, 64])]; + tensor var_954_cast_fp16 = reshape(shape = var_953, x = tensor_59_cast_fp16)[name = tensor("op_954_cast_fp16")]; + tensor var_955_perm_0 = const()[name = tensor("op_955_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_957 = const()[name = tensor("op_957"), val = tensor([20, -1, 64])]; + tensor var_955_cast_fp16 = transpose(perm = var_955_perm_0, x = var_954_cast_fp16)[name = tensor("transpose_90")]; + tensor query_states_19_cast_fp16 = reshape(shape = var_957, x = var_955_cast_fp16)[name = tensor("query_states_19_cast_fp16")]; + tensor var_959 = const()[name = tensor("op_959"), val = tensor([20, -1, 64])]; + tensor var_939_cast_fp16 = transpose(perm = var_939_perm_0, x = var_938_cast_fp16)[name = tensor("transpose_92")]; + tensor key_states_39_cast_fp16 = reshape(shape = var_959, x = var_939_cast_fp16)[name = tensor("key_states_39_cast_fp16")]; + tensor var_961 = const()[name = tensor("op_961"), val = tensor([20, -1, 64])]; + tensor var_946_cast_fp16 = transpose(perm = var_946_perm_0, x = var_945_cast_fp16)[name = tensor("transpose_91")]; + tensor value_states_39_cast_fp16 = reshape(shape = var_961, x = var_946_cast_fp16)[name = tensor("value_states_39_cast_fp16")]; + tensor attn_weights_55_transpose_x_1 = const()[name = tensor("attn_weights_55_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_55_transpose_y_1 = const()[name = tensor("attn_weights_55_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_55_cast_fp16 = matmul(transpose_x = attn_weights_55_transpose_x_1, transpose_y = attn_weights_55_transpose_y_1, x = query_states_19_cast_fp16, y = key_states_39_cast_fp16)[name = tensor("attn_weights_55_cast_fp16")]; + tensor var_966 = const()[name = tensor("op_966"), val = tensor([1, 20, 77, 77])]; + tensor var_967_cast_fp16 = reshape(shape = var_966, x = attn_weights_55_cast_fp16)[name = tensor("op_967_cast_fp16")]; + tensor attn_weights_57_cast_fp16 = add(x = var_967_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_57_cast_fp16")]; + tensor var_972 = const()[name = tensor("op_972"), val = tensor([20, 77, 77])]; + tensor input_149_cast_fp16 = reshape(shape = var_972, x = attn_weights_57_cast_fp16)[name = tensor("input_149_cast_fp16")]; + tensor input_151_cast_fp16 = softmax(axis = var_5, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; + tensor attn_output_55_transpose_x_0 = const()[name = tensor("attn_output_55_transpose_x_0"), val = tensor(false)]; + tensor attn_output_55_transpose_y_0 = const()[name = tensor("attn_output_55_transpose_y_0"), val = tensor(false)]; + tensor attn_output_55_cast_fp16 = matmul(transpose_x = attn_output_55_transpose_x_0, transpose_y = attn_output_55_transpose_y_0, x = input_151_cast_fp16, y = value_states_39_cast_fp16)[name = tensor("attn_output_55_cast_fp16")]; + tensor var_977 = const()[name = tensor("op_977"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_57_cast_fp16 = reshape(shape = var_977, x = attn_output_55_cast_fp16)[name = tensor("attn_output_57_cast_fp16")]; + tensor attn_output_59_perm_0 = const()[name = tensor("attn_output_59_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_980 = const()[name = tensor("op_980"), val = tensor([1, 77, 1280])]; + tensor attn_output_59_cast_fp16 = transpose(perm = attn_output_59_perm_0, x = attn_output_57_cast_fp16)[name = tensor("transpose_89")]; + tensor input_153_cast_fp16 = reshape(shape = var_980, x = attn_output_59_cast_fp16)[name = tensor("input_153_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(490740608)))]; + tensor text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494017472)))]; + tensor linear_57_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16, x = input_153_cast_fp16)[name = tensor("linear_57_cast_fp16")]; + tensor input_155_cast_fp16 = add(x = input_147_cast_fp16, y = linear_57_cast_fp16)[name = tensor("input_155_cast_fp16")]; + tensor input_157_axes_0 = const()[name = tensor("input_157_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494020096)))]; + tensor text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494022720)))]; + tensor input_157_cast_fp16 = layer_norm(axes = input_157_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16, x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(494025344)))]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507132608)))]; + tensor linear_58_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("linear_58_cast_fp16")]; + tensor input_161_mode_0 = const()[name = tensor("input_161_mode_0"), val = tensor("EXACT")]; + tensor input_161_cast_fp16 = gelu(mode = input_161_mode_0, x = linear_58_cast_fp16)[name = tensor("input_161_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(507142912)))]; + tensor text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520250176)))]; + tensor linear_59_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("linear_59_cast_fp16")]; + tensor input_163_cast_fp16 = add(x = input_155_cast_fp16, y = linear_59_cast_fp16)[name = tensor("input_163_cast_fp16")]; + tensor hidden_states_61_axes_0 = const()[name = tensor("hidden_states_61_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520252800)))]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520255424)))]; + tensor hidden_states_61_cast_fp16 = layer_norm(axes = hidden_states_61_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("hidden_states_61_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(520258048)))]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523534912)))]; + tensor linear_60_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor("linear_60_cast_fp16")]; + tensor var_1019_to_fp16 = const()[name = tensor("op_1019_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_65_cast_fp16 = mul(x = linear_60_cast_fp16, y = var_1019_to_fp16)[name = tensor("tensor_65_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(523537536)))]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526814400)))]; + tensor linear_61_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor("linear_61_cast_fp16")]; + tensor var_1024 = const()[name = tensor("op_1024"), val = tensor([1, -1, 20, 64])]; + tensor var_1025_cast_fp16 = reshape(shape = var_1024, x = linear_61_cast_fp16)[name = tensor("op_1025_cast_fp16")]; + tensor var_1026_perm_0 = const()[name = tensor("op_1026_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526817024)))]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530093888)))]; + tensor linear_62_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor("linear_62_cast_fp16")]; + tensor var_1031 = const()[name = tensor("op_1031"), val = tensor([1, -1, 20, 64])]; + tensor var_1032_cast_fp16 = reshape(shape = var_1031, x = linear_62_cast_fp16)[name = tensor("op_1032_cast_fp16")]; + tensor var_1033_perm_0 = const()[name = tensor("op_1033_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1040 = const()[name = tensor("op_1040"), val = tensor([1, 77, 20, 64])]; + tensor var_1041_cast_fp16 = reshape(shape = var_1040, x = tensor_65_cast_fp16)[name = tensor("op_1041_cast_fp16")]; + tensor var_1042_perm_0 = const()[name = tensor("op_1042_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1044 = const()[name = tensor("op_1044"), val = tensor([20, -1, 64])]; + tensor var_1042_cast_fp16 = transpose(perm = var_1042_perm_0, x = var_1041_cast_fp16)[name = tensor("transpose_86")]; + tensor query_states_21_cast_fp16 = reshape(shape = var_1044, x = var_1042_cast_fp16)[name = tensor("query_states_21_cast_fp16")]; + tensor var_1046 = const()[name = tensor("op_1046"), val = tensor([20, -1, 64])]; + tensor var_1026_cast_fp16 = transpose(perm = var_1026_perm_0, x = var_1025_cast_fp16)[name = tensor("transpose_88")]; + tensor key_states_43_cast_fp16 = reshape(shape = var_1046, x = var_1026_cast_fp16)[name = tensor("key_states_43_cast_fp16")]; + tensor var_1048 = const()[name = tensor("op_1048"), val = tensor([20, -1, 64])]; + tensor var_1033_cast_fp16 = transpose(perm = var_1033_perm_0, x = var_1032_cast_fp16)[name = tensor("transpose_87")]; + tensor value_states_43_cast_fp16 = reshape(shape = var_1048, x = var_1033_cast_fp16)[name = tensor("value_states_43_cast_fp16")]; + tensor attn_weights_61_transpose_x_1 = const()[name = tensor("attn_weights_61_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_61_transpose_y_1 = const()[name = tensor("attn_weights_61_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_61_cast_fp16 = matmul(transpose_x = attn_weights_61_transpose_x_1, transpose_y = attn_weights_61_transpose_y_1, x = query_states_21_cast_fp16, y = key_states_43_cast_fp16)[name = tensor("attn_weights_61_cast_fp16")]; + tensor var_1053 = const()[name = tensor("op_1053"), val = tensor([1, 20, 77, 77])]; + tensor var_1054_cast_fp16 = reshape(shape = var_1053, x = attn_weights_61_cast_fp16)[name = tensor("op_1054_cast_fp16")]; + tensor attn_weights_63_cast_fp16 = add(x = var_1054_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_63_cast_fp16")]; + tensor var_1059 = const()[name = tensor("op_1059"), val = tensor([20, 77, 77])]; + tensor input_165_cast_fp16 = reshape(shape = var_1059, x = attn_weights_63_cast_fp16)[name = tensor("input_165_cast_fp16")]; + tensor input_167_cast_fp16 = softmax(axis = var_5, x = input_165_cast_fp16)[name = tensor("input_167_cast_fp16")]; + tensor attn_output_61_transpose_x_0 = const()[name = tensor("attn_output_61_transpose_x_0"), val = tensor(false)]; + tensor attn_output_61_transpose_y_0 = const()[name = tensor("attn_output_61_transpose_y_0"), val = tensor(false)]; + tensor attn_output_61_cast_fp16 = matmul(transpose_x = attn_output_61_transpose_x_0, transpose_y = attn_output_61_transpose_y_0, x = input_167_cast_fp16, y = value_states_43_cast_fp16)[name = tensor("attn_output_61_cast_fp16")]; + tensor var_1064 = const()[name = tensor("op_1064"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_63_cast_fp16 = reshape(shape = var_1064, x = attn_output_61_cast_fp16)[name = tensor("attn_output_63_cast_fp16")]; + tensor attn_output_65_perm_0 = const()[name = tensor("attn_output_65_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1067 = const()[name = tensor("op_1067"), val = tensor([1, 77, 1280])]; + tensor attn_output_65_cast_fp16 = transpose(perm = attn_output_65_perm_0, x = attn_output_63_cast_fp16)[name = tensor("transpose_85")]; + tensor input_169_cast_fp16 = reshape(shape = var_1067, x = attn_output_65_cast_fp16)[name = tensor("input_169_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(530096512)))]; + tensor text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533373376)))]; + tensor linear_63_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16, x = input_169_cast_fp16)[name = tensor("linear_63_cast_fp16")]; + tensor input_171_cast_fp16 = add(x = input_163_cast_fp16, y = linear_63_cast_fp16)[name = tensor("input_171_cast_fp16")]; + tensor input_173_axes_0 = const()[name = tensor("input_173_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533376000)))]; + tensor text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533378624)))]; + tensor input_173_cast_fp16 = layer_norm(axes = input_173_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16, x = input_171_cast_fp16)[name = tensor("input_173_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(533381248)))]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(546488512)))]; + tensor linear_64_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("linear_64_cast_fp16")]; + tensor input_177_mode_0 = const()[name = tensor("input_177_mode_0"), val = tensor("EXACT")]; + tensor input_177_cast_fp16 = gelu(mode = input_177_mode_0, x = linear_64_cast_fp16)[name = tensor("input_177_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(546498816)))]; + tensor text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559606080)))]; + tensor linear_65_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("linear_65_cast_fp16")]; + tensor input_179_cast_fp16 = add(x = input_171_cast_fp16, y = linear_65_cast_fp16)[name = tensor("input_179_cast_fp16")]; + tensor hidden_states_67_axes_0 = const()[name = tensor("hidden_states_67_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559608704)))]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559611328)))]; + tensor hidden_states_67_cast_fp16 = layer_norm(axes = hidden_states_67_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16, x = input_179_cast_fp16)[name = tensor("hidden_states_67_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(559613952)))]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(562890816)))]; + tensor linear_66_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor("linear_66_cast_fp16")]; + tensor var_1106_to_fp16 = const()[name = tensor("op_1106_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_71_cast_fp16 = mul(x = linear_66_cast_fp16, y = var_1106_to_fp16)[name = tensor("tensor_71_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(562893440)))]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(566170304)))]; + tensor linear_67_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor("linear_67_cast_fp16")]; + tensor var_1111 = const()[name = tensor("op_1111"), val = tensor([1, -1, 20, 64])]; + tensor var_1112_cast_fp16 = reshape(shape = var_1111, x = linear_67_cast_fp16)[name = tensor("op_1112_cast_fp16")]; + tensor var_1113_perm_0 = const()[name = tensor("op_1113_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(566172928)))]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569449792)))]; + tensor linear_68_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor("linear_68_cast_fp16")]; + tensor var_1118 = const()[name = tensor("op_1118"), val = tensor([1, -1, 20, 64])]; + tensor var_1119_cast_fp16 = reshape(shape = var_1118, x = linear_68_cast_fp16)[name = tensor("op_1119_cast_fp16")]; + tensor var_1120_perm_0 = const()[name = tensor("op_1120_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1127 = const()[name = tensor("op_1127"), val = tensor([1, 77, 20, 64])]; + tensor var_1128_cast_fp16 = reshape(shape = var_1127, x = tensor_71_cast_fp16)[name = tensor("op_1128_cast_fp16")]; + tensor var_1129_perm_0 = const()[name = tensor("op_1129_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1131 = const()[name = tensor("op_1131"), val = tensor([20, -1, 64])]; + tensor var_1129_cast_fp16 = transpose(perm = var_1129_perm_0, x = var_1128_cast_fp16)[name = tensor("transpose_82")]; + tensor query_states_23_cast_fp16 = reshape(shape = var_1131, x = var_1129_cast_fp16)[name = tensor("query_states_23_cast_fp16")]; + tensor var_1133 = const()[name = tensor("op_1133"), val = tensor([20, -1, 64])]; + tensor var_1113_cast_fp16 = transpose(perm = var_1113_perm_0, x = var_1112_cast_fp16)[name = tensor("transpose_84")]; + tensor key_states_47_cast_fp16 = reshape(shape = var_1133, x = var_1113_cast_fp16)[name = tensor("key_states_47_cast_fp16")]; + tensor var_1135 = const()[name = tensor("op_1135"), val = tensor([20, -1, 64])]; + tensor var_1120_cast_fp16 = transpose(perm = var_1120_perm_0, x = var_1119_cast_fp16)[name = tensor("transpose_83")]; + tensor value_states_47_cast_fp16 = reshape(shape = var_1135, x = var_1120_cast_fp16)[name = tensor("value_states_47_cast_fp16")]; + tensor attn_weights_67_transpose_x_1 = const()[name = tensor("attn_weights_67_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_67_transpose_y_1 = const()[name = tensor("attn_weights_67_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_67_cast_fp16 = matmul(transpose_x = attn_weights_67_transpose_x_1, transpose_y = attn_weights_67_transpose_y_1, x = query_states_23_cast_fp16, y = key_states_47_cast_fp16)[name = tensor("attn_weights_67_cast_fp16")]; + tensor var_1140 = const()[name = tensor("op_1140"), val = tensor([1, 20, 77, 77])]; + tensor var_1141_cast_fp16 = reshape(shape = var_1140, x = attn_weights_67_cast_fp16)[name = tensor("op_1141_cast_fp16")]; + tensor attn_weights_69_cast_fp16 = add(x = var_1141_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_69_cast_fp16")]; + tensor var_1146 = const()[name = tensor("op_1146"), val = tensor([20, 77, 77])]; + tensor input_181_cast_fp16 = reshape(shape = var_1146, x = attn_weights_69_cast_fp16)[name = tensor("input_181_cast_fp16")]; + tensor input_183_cast_fp16 = softmax(axis = var_5, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; + tensor attn_output_67_transpose_x_0 = const()[name = tensor("attn_output_67_transpose_x_0"), val = tensor(false)]; + tensor attn_output_67_transpose_y_0 = const()[name = tensor("attn_output_67_transpose_y_0"), val = tensor(false)]; + tensor attn_output_67_cast_fp16 = matmul(transpose_x = attn_output_67_transpose_x_0, transpose_y = attn_output_67_transpose_y_0, x = input_183_cast_fp16, y = value_states_47_cast_fp16)[name = tensor("attn_output_67_cast_fp16")]; + tensor var_1151 = const()[name = tensor("op_1151"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_69_cast_fp16 = reshape(shape = var_1151, x = attn_output_67_cast_fp16)[name = tensor("attn_output_69_cast_fp16")]; + tensor attn_output_71_perm_0 = const()[name = tensor("attn_output_71_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1154 = const()[name = tensor("op_1154"), val = tensor([1, 77, 1280])]; + tensor attn_output_71_cast_fp16 = transpose(perm = attn_output_71_perm_0, x = attn_output_69_cast_fp16)[name = tensor("transpose_81")]; + tensor input_185_cast_fp16 = reshape(shape = var_1154, x = attn_output_71_cast_fp16)[name = tensor("input_185_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(569452416)))]; + tensor text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572729280)))]; + tensor linear_69_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16, x = input_185_cast_fp16)[name = tensor("linear_69_cast_fp16")]; + tensor input_187_cast_fp16 = add(x = input_179_cast_fp16, y = linear_69_cast_fp16)[name = tensor("input_187_cast_fp16")]; + tensor input_189_axes_0 = const()[name = tensor("input_189_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572731904)))]; + tensor text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572734528)))]; + tensor input_189_cast_fp16 = layer_norm(axes = input_189_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("input_189_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(572737152)))]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585844416)))]; + tensor linear_70_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16, x = input_189_cast_fp16)[name = tensor("linear_70_cast_fp16")]; + tensor input_193_mode_0 = const()[name = tensor("input_193_mode_0"), val = tensor("EXACT")]; + tensor input_193_cast_fp16 = gelu(mode = input_193_mode_0, x = linear_70_cast_fp16)[name = tensor("input_193_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(585854720)))]; + tensor text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(598961984)))]; + tensor linear_71_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16, x = input_193_cast_fp16)[name = tensor("linear_71_cast_fp16")]; + tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_71_cast_fp16)[name = tensor("input_195_cast_fp16")]; + tensor hidden_states_73_axes_0 = const()[name = tensor("hidden_states_73_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(598964608)))]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(598967232)))]; + tensor hidden_states_73_cast_fp16 = layer_norm(axes = hidden_states_73_axes_0, beta = text_encoder_text_model_encoder_layers_12_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_12_layer_norm1_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("hidden_states_73_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(598969856)))]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(602246720)))]; + tensor linear_72_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_q_proj_weight_to_fp16, x = hidden_states_73_cast_fp16)[name = tensor("linear_72_cast_fp16")]; + tensor var_1193_to_fp16 = const()[name = tensor("op_1193_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_77_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1193_to_fp16)[name = tensor("tensor_77_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(602249344)))]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(605526208)))]; + tensor linear_73_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_k_proj_weight_to_fp16, x = hidden_states_73_cast_fp16)[name = tensor("linear_73_cast_fp16")]; + tensor var_1198 = const()[name = tensor("op_1198"), val = tensor([1, -1, 20, 64])]; + tensor var_1199_cast_fp16 = reshape(shape = var_1198, x = linear_73_cast_fp16)[name = tensor("op_1199_cast_fp16")]; + tensor var_1200_perm_0 = const()[name = tensor("op_1200_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(605528832)))]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(608805696)))]; + tensor linear_74_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_v_proj_weight_to_fp16, x = hidden_states_73_cast_fp16)[name = tensor("linear_74_cast_fp16")]; + tensor var_1205 = const()[name = tensor("op_1205"), val = tensor([1, -1, 20, 64])]; + tensor var_1206_cast_fp16 = reshape(shape = var_1205, x = linear_74_cast_fp16)[name = tensor("op_1206_cast_fp16")]; + tensor var_1207_perm_0 = const()[name = tensor("op_1207_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1214 = const()[name = tensor("op_1214"), val = tensor([1, 77, 20, 64])]; + tensor var_1215_cast_fp16 = reshape(shape = var_1214, x = tensor_77_cast_fp16)[name = tensor("op_1215_cast_fp16")]; + tensor var_1216_perm_0 = const()[name = tensor("op_1216_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1218 = const()[name = tensor("op_1218"), val = tensor([20, -1, 64])]; + tensor var_1216_cast_fp16 = transpose(perm = var_1216_perm_0, x = var_1215_cast_fp16)[name = tensor("transpose_78")]; + tensor query_states_25_cast_fp16 = reshape(shape = var_1218, x = var_1216_cast_fp16)[name = tensor("query_states_25_cast_fp16")]; + tensor var_1220 = const()[name = tensor("op_1220"), val = tensor([20, -1, 64])]; + tensor var_1200_cast_fp16 = transpose(perm = var_1200_perm_0, x = var_1199_cast_fp16)[name = tensor("transpose_80")]; + tensor key_states_51_cast_fp16 = reshape(shape = var_1220, x = var_1200_cast_fp16)[name = tensor("key_states_51_cast_fp16")]; + tensor var_1222 = const()[name = tensor("op_1222"), val = tensor([20, -1, 64])]; + tensor var_1207_cast_fp16 = transpose(perm = var_1207_perm_0, x = var_1206_cast_fp16)[name = tensor("transpose_79")]; + tensor value_states_51_cast_fp16 = reshape(shape = var_1222, x = var_1207_cast_fp16)[name = tensor("value_states_51_cast_fp16")]; + tensor attn_weights_73_transpose_x_1 = const()[name = tensor("attn_weights_73_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_73_transpose_y_1 = const()[name = tensor("attn_weights_73_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_73_cast_fp16 = matmul(transpose_x = attn_weights_73_transpose_x_1, transpose_y = attn_weights_73_transpose_y_1, x = query_states_25_cast_fp16, y = key_states_51_cast_fp16)[name = tensor("attn_weights_73_cast_fp16")]; + tensor var_1227 = const()[name = tensor("op_1227"), val = tensor([1, 20, 77, 77])]; + tensor var_1228_cast_fp16 = reshape(shape = var_1227, x = attn_weights_73_cast_fp16)[name = tensor("op_1228_cast_fp16")]; + tensor attn_weights_75_cast_fp16 = add(x = var_1228_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_75_cast_fp16")]; + tensor var_1233 = const()[name = tensor("op_1233"), val = tensor([20, 77, 77])]; + tensor input_197_cast_fp16 = reshape(shape = var_1233, x = attn_weights_75_cast_fp16)[name = tensor("input_197_cast_fp16")]; + tensor input_199_cast_fp16 = softmax(axis = var_5, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; + tensor attn_output_73_transpose_x_0 = const()[name = tensor("attn_output_73_transpose_x_0"), val = tensor(false)]; + tensor attn_output_73_transpose_y_0 = const()[name = tensor("attn_output_73_transpose_y_0"), val = tensor(false)]; + tensor attn_output_73_cast_fp16 = matmul(transpose_x = attn_output_73_transpose_x_0, transpose_y = attn_output_73_transpose_y_0, x = input_199_cast_fp16, y = value_states_51_cast_fp16)[name = tensor("attn_output_73_cast_fp16")]; + tensor var_1238 = const()[name = tensor("op_1238"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_75_cast_fp16 = reshape(shape = var_1238, x = attn_output_73_cast_fp16)[name = tensor("attn_output_75_cast_fp16")]; + tensor attn_output_77_perm_0 = const()[name = tensor("attn_output_77_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1241 = const()[name = tensor("op_1241"), val = tensor([1, 77, 1280])]; + tensor attn_output_77_cast_fp16 = transpose(perm = attn_output_77_perm_0, x = attn_output_75_cast_fp16)[name = tensor("transpose_77")]; + tensor input_201_cast_fp16 = reshape(shape = var_1241, x = attn_output_77_cast_fp16)[name = tensor("input_201_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(608808320)))]; + tensor text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(612085184)))]; + tensor linear_75_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_self_attn_out_proj_weight_to_fp16, x = input_201_cast_fp16)[name = tensor("linear_75_cast_fp16")]; + tensor input_203_cast_fp16 = add(x = input_195_cast_fp16, y = linear_75_cast_fp16)[name = tensor("input_203_cast_fp16")]; + tensor input_205_axes_0 = const()[name = tensor("input_205_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(612087808)))]; + tensor text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(612090432)))]; + tensor input_205_cast_fp16 = layer_norm(axes = input_205_axes_0, beta = text_encoder_text_model_encoder_layers_12_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_12_layer_norm2_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("input_205_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(612093056)))]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(625200320)))]; + tensor linear_76_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_mlp_fc1_weight_to_fp16, x = input_205_cast_fp16)[name = tensor("linear_76_cast_fp16")]; + tensor input_209_mode_0 = const()[name = tensor("input_209_mode_0"), val = tensor("EXACT")]; + tensor input_209_cast_fp16 = gelu(mode = input_209_mode_0, x = linear_76_cast_fp16)[name = tensor("input_209_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(625210624)))]; + tensor text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(638317888)))]; + tensor linear_77_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_12_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_12_mlp_fc2_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("linear_77_cast_fp16")]; + tensor input_211_cast_fp16 = add(x = input_203_cast_fp16, y = linear_77_cast_fp16)[name = tensor("input_211_cast_fp16")]; + tensor hidden_states_79_axes_0 = const()[name = tensor("hidden_states_79_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(638320512)))]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(638323136)))]; + tensor hidden_states_79_cast_fp16 = layer_norm(axes = hidden_states_79_axes_0, beta = text_encoder_text_model_encoder_layers_13_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_13_layer_norm1_weight_to_fp16, x = input_211_cast_fp16)[name = tensor("hidden_states_79_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(638325760)))]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(641602624)))]; + tensor linear_78_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_q_proj_weight_to_fp16, x = hidden_states_79_cast_fp16)[name = tensor("linear_78_cast_fp16")]; + tensor var_1280_to_fp16 = const()[name = tensor("op_1280_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_83_cast_fp16 = mul(x = linear_78_cast_fp16, y = var_1280_to_fp16)[name = tensor("tensor_83_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(641605248)))]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(644882112)))]; + tensor linear_79_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_k_proj_weight_to_fp16, x = hidden_states_79_cast_fp16)[name = tensor("linear_79_cast_fp16")]; + tensor var_1285 = const()[name = tensor("op_1285"), val = tensor([1, -1, 20, 64])]; + tensor var_1286_cast_fp16 = reshape(shape = var_1285, x = linear_79_cast_fp16)[name = tensor("op_1286_cast_fp16")]; + tensor var_1287_perm_0 = const()[name = tensor("op_1287_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(644884736)))]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(648161600)))]; + tensor linear_80_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_v_proj_weight_to_fp16, x = hidden_states_79_cast_fp16)[name = tensor("linear_80_cast_fp16")]; + tensor var_1292 = const()[name = tensor("op_1292"), val = tensor([1, -1, 20, 64])]; + tensor var_1293_cast_fp16 = reshape(shape = var_1292, x = linear_80_cast_fp16)[name = tensor("op_1293_cast_fp16")]; + tensor var_1294_perm_0 = const()[name = tensor("op_1294_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1301 = const()[name = tensor("op_1301"), val = tensor([1, 77, 20, 64])]; + tensor var_1302_cast_fp16 = reshape(shape = var_1301, x = tensor_83_cast_fp16)[name = tensor("op_1302_cast_fp16")]; + tensor var_1303_perm_0 = const()[name = tensor("op_1303_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1305 = const()[name = tensor("op_1305"), val = tensor([20, -1, 64])]; + tensor var_1303_cast_fp16 = transpose(perm = var_1303_perm_0, x = var_1302_cast_fp16)[name = tensor("transpose_74")]; + tensor query_states_27_cast_fp16 = reshape(shape = var_1305, x = var_1303_cast_fp16)[name = tensor("query_states_27_cast_fp16")]; + tensor var_1307 = const()[name = tensor("op_1307"), val = tensor([20, -1, 64])]; + tensor var_1287_cast_fp16 = transpose(perm = var_1287_perm_0, x = var_1286_cast_fp16)[name = tensor("transpose_76")]; + tensor key_states_55_cast_fp16 = reshape(shape = var_1307, x = var_1287_cast_fp16)[name = tensor("key_states_55_cast_fp16")]; + tensor var_1309 = const()[name = tensor("op_1309"), val = tensor([20, -1, 64])]; + tensor var_1294_cast_fp16 = transpose(perm = var_1294_perm_0, x = var_1293_cast_fp16)[name = tensor("transpose_75")]; + tensor value_states_55_cast_fp16 = reshape(shape = var_1309, x = var_1294_cast_fp16)[name = tensor("value_states_55_cast_fp16")]; + tensor attn_weights_79_transpose_x_1 = const()[name = tensor("attn_weights_79_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_79_transpose_y_1 = const()[name = tensor("attn_weights_79_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_79_cast_fp16 = matmul(transpose_x = attn_weights_79_transpose_x_1, transpose_y = attn_weights_79_transpose_y_1, x = query_states_27_cast_fp16, y = key_states_55_cast_fp16)[name = tensor("attn_weights_79_cast_fp16")]; + tensor var_1314 = const()[name = tensor("op_1314"), val = tensor([1, 20, 77, 77])]; + tensor var_1315_cast_fp16 = reshape(shape = var_1314, x = attn_weights_79_cast_fp16)[name = tensor("op_1315_cast_fp16")]; + tensor attn_weights_81_cast_fp16 = add(x = var_1315_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_81_cast_fp16")]; + tensor var_1320 = const()[name = tensor("op_1320"), val = tensor([20, 77, 77])]; + tensor input_213_cast_fp16 = reshape(shape = var_1320, x = attn_weights_81_cast_fp16)[name = tensor("input_213_cast_fp16")]; + tensor input_215_cast_fp16 = softmax(axis = var_5, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; + tensor attn_output_79_transpose_x_0 = const()[name = tensor("attn_output_79_transpose_x_0"), val = tensor(false)]; + tensor attn_output_79_transpose_y_0 = const()[name = tensor("attn_output_79_transpose_y_0"), val = tensor(false)]; + tensor attn_output_79_cast_fp16 = matmul(transpose_x = attn_output_79_transpose_x_0, transpose_y = attn_output_79_transpose_y_0, x = input_215_cast_fp16, y = value_states_55_cast_fp16)[name = tensor("attn_output_79_cast_fp16")]; + tensor var_1325 = const()[name = tensor("op_1325"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_81_cast_fp16 = reshape(shape = var_1325, x = attn_output_79_cast_fp16)[name = tensor("attn_output_81_cast_fp16")]; + tensor attn_output_83_perm_0 = const()[name = tensor("attn_output_83_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1328 = const()[name = tensor("op_1328"), val = tensor([1, 77, 1280])]; + tensor attn_output_83_cast_fp16 = transpose(perm = attn_output_83_perm_0, x = attn_output_81_cast_fp16)[name = tensor("transpose_73")]; + tensor input_217_cast_fp16 = reshape(shape = var_1328, x = attn_output_83_cast_fp16)[name = tensor("input_217_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(648164224)))]; + tensor text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(651441088)))]; + tensor linear_81_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_self_attn_out_proj_weight_to_fp16, x = input_217_cast_fp16)[name = tensor("linear_81_cast_fp16")]; + tensor input_219_cast_fp16 = add(x = input_211_cast_fp16, y = linear_81_cast_fp16)[name = tensor("input_219_cast_fp16")]; + tensor input_221_axes_0 = const()[name = tensor("input_221_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(651443712)))]; + tensor text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(651446336)))]; + tensor input_221_cast_fp16 = layer_norm(axes = input_221_axes_0, beta = text_encoder_text_model_encoder_layers_13_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_13_layer_norm2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(651448960)))]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(664556224)))]; + tensor linear_82_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_mlp_fc1_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("linear_82_cast_fp16")]; + tensor input_225_mode_0 = const()[name = tensor("input_225_mode_0"), val = tensor("EXACT")]; + tensor input_225_cast_fp16 = gelu(mode = input_225_mode_0, x = linear_82_cast_fp16)[name = tensor("input_225_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(664566528)))]; + tensor text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(677673792)))]; + tensor linear_83_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_13_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_13_mlp_fc2_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("linear_83_cast_fp16")]; + tensor input_227_cast_fp16 = add(x = input_219_cast_fp16, y = linear_83_cast_fp16)[name = tensor("input_227_cast_fp16")]; + tensor hidden_states_85_axes_0 = const()[name = tensor("hidden_states_85_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(677676416)))]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(677679040)))]; + tensor hidden_states_85_cast_fp16 = layer_norm(axes = hidden_states_85_axes_0, beta = text_encoder_text_model_encoder_layers_14_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_14_layer_norm1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("hidden_states_85_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(677681664)))]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(680958528)))]; + tensor linear_84_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_q_proj_weight_to_fp16, x = hidden_states_85_cast_fp16)[name = tensor("linear_84_cast_fp16")]; + tensor var_1367_to_fp16 = const()[name = tensor("op_1367_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_89_cast_fp16 = mul(x = linear_84_cast_fp16, y = var_1367_to_fp16)[name = tensor("tensor_89_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(680961152)))]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(684238016)))]; + tensor linear_85_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_k_proj_weight_to_fp16, x = hidden_states_85_cast_fp16)[name = tensor("linear_85_cast_fp16")]; + tensor var_1372 = const()[name = tensor("op_1372"), val = tensor([1, -1, 20, 64])]; + tensor var_1373_cast_fp16 = reshape(shape = var_1372, x = linear_85_cast_fp16)[name = tensor("op_1373_cast_fp16")]; + tensor var_1374_perm_0 = const()[name = tensor("op_1374_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(684240640)))]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(687517504)))]; + tensor linear_86_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_v_proj_weight_to_fp16, x = hidden_states_85_cast_fp16)[name = tensor("linear_86_cast_fp16")]; + tensor var_1379 = const()[name = tensor("op_1379"), val = tensor([1, -1, 20, 64])]; + tensor var_1380_cast_fp16 = reshape(shape = var_1379, x = linear_86_cast_fp16)[name = tensor("op_1380_cast_fp16")]; + tensor var_1381_perm_0 = const()[name = tensor("op_1381_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1388 = const()[name = tensor("op_1388"), val = tensor([1, 77, 20, 64])]; + tensor var_1389_cast_fp16 = reshape(shape = var_1388, x = tensor_89_cast_fp16)[name = tensor("op_1389_cast_fp16")]; + tensor var_1390_perm_0 = const()[name = tensor("op_1390_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1392 = const()[name = tensor("op_1392"), val = tensor([20, -1, 64])]; + tensor var_1390_cast_fp16 = transpose(perm = var_1390_perm_0, x = var_1389_cast_fp16)[name = tensor("transpose_70")]; + tensor query_states_29_cast_fp16 = reshape(shape = var_1392, x = var_1390_cast_fp16)[name = tensor("query_states_29_cast_fp16")]; + tensor var_1394 = const()[name = tensor("op_1394"), val = tensor([20, -1, 64])]; + tensor var_1374_cast_fp16 = transpose(perm = var_1374_perm_0, x = var_1373_cast_fp16)[name = tensor("transpose_72")]; + tensor key_states_59_cast_fp16 = reshape(shape = var_1394, x = var_1374_cast_fp16)[name = tensor("key_states_59_cast_fp16")]; + tensor var_1396 = const()[name = tensor("op_1396"), val = tensor([20, -1, 64])]; + tensor var_1381_cast_fp16 = transpose(perm = var_1381_perm_0, x = var_1380_cast_fp16)[name = tensor("transpose_71")]; + tensor value_states_59_cast_fp16 = reshape(shape = var_1396, x = var_1381_cast_fp16)[name = tensor("value_states_59_cast_fp16")]; + tensor attn_weights_85_transpose_x_1 = const()[name = tensor("attn_weights_85_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_85_transpose_y_1 = const()[name = tensor("attn_weights_85_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_85_cast_fp16 = matmul(transpose_x = attn_weights_85_transpose_x_1, transpose_y = attn_weights_85_transpose_y_1, x = query_states_29_cast_fp16, y = key_states_59_cast_fp16)[name = tensor("attn_weights_85_cast_fp16")]; + tensor var_1401 = const()[name = tensor("op_1401"), val = tensor([1, 20, 77, 77])]; + tensor var_1402_cast_fp16 = reshape(shape = var_1401, x = attn_weights_85_cast_fp16)[name = tensor("op_1402_cast_fp16")]; + tensor attn_weights_87_cast_fp16 = add(x = var_1402_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_87_cast_fp16")]; + tensor var_1407 = const()[name = tensor("op_1407"), val = tensor([20, 77, 77])]; + tensor input_229_cast_fp16 = reshape(shape = var_1407, x = attn_weights_87_cast_fp16)[name = tensor("input_229_cast_fp16")]; + tensor input_231_cast_fp16 = softmax(axis = var_5, x = input_229_cast_fp16)[name = tensor("input_231_cast_fp16")]; + tensor attn_output_85_transpose_x_0 = const()[name = tensor("attn_output_85_transpose_x_0"), val = tensor(false)]; + tensor attn_output_85_transpose_y_0 = const()[name = tensor("attn_output_85_transpose_y_0"), val = tensor(false)]; + tensor attn_output_85_cast_fp16 = matmul(transpose_x = attn_output_85_transpose_x_0, transpose_y = attn_output_85_transpose_y_0, x = input_231_cast_fp16, y = value_states_59_cast_fp16)[name = tensor("attn_output_85_cast_fp16")]; + tensor var_1412 = const()[name = tensor("op_1412"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_87_cast_fp16 = reshape(shape = var_1412, x = attn_output_85_cast_fp16)[name = tensor("attn_output_87_cast_fp16")]; + tensor attn_output_89_perm_0 = const()[name = tensor("attn_output_89_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1415 = const()[name = tensor("op_1415"), val = tensor([1, 77, 1280])]; + tensor attn_output_89_cast_fp16 = transpose(perm = attn_output_89_perm_0, x = attn_output_87_cast_fp16)[name = tensor("transpose_69")]; + tensor input_233_cast_fp16 = reshape(shape = var_1415, x = attn_output_89_cast_fp16)[name = tensor("input_233_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(687520128)))]; + tensor text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(690796992)))]; + tensor linear_87_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_self_attn_out_proj_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("linear_87_cast_fp16")]; + tensor input_235_cast_fp16 = add(x = input_227_cast_fp16, y = linear_87_cast_fp16)[name = tensor("input_235_cast_fp16")]; + tensor input_237_axes_0 = const()[name = tensor("input_237_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(690799616)))]; + tensor text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(690802240)))]; + tensor input_237_cast_fp16 = layer_norm(axes = input_237_axes_0, beta = text_encoder_text_model_encoder_layers_14_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_14_layer_norm2_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("input_237_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(690804864)))]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(703912128)))]; + tensor linear_88_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_mlp_fc1_weight_to_fp16, x = input_237_cast_fp16)[name = tensor("linear_88_cast_fp16")]; + tensor input_241_mode_0 = const()[name = tensor("input_241_mode_0"), val = tensor("EXACT")]; + tensor input_241_cast_fp16 = gelu(mode = input_241_mode_0, x = linear_88_cast_fp16)[name = tensor("input_241_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(703922432)))]; + tensor text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(717029696)))]; + tensor linear_89_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_14_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_14_mlp_fc2_weight_to_fp16, x = input_241_cast_fp16)[name = tensor("linear_89_cast_fp16")]; + tensor input_243_cast_fp16 = add(x = input_235_cast_fp16, y = linear_89_cast_fp16)[name = tensor("input_243_cast_fp16")]; + tensor hidden_states_91_axes_0 = const()[name = tensor("hidden_states_91_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(717032320)))]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(717034944)))]; + tensor hidden_states_91_cast_fp16 = layer_norm(axes = hidden_states_91_axes_0, beta = text_encoder_text_model_encoder_layers_15_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_15_layer_norm1_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("hidden_states_91_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(717037568)))]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(720314432)))]; + tensor linear_90_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_q_proj_weight_to_fp16, x = hidden_states_91_cast_fp16)[name = tensor("linear_90_cast_fp16")]; + tensor var_1454_to_fp16 = const()[name = tensor("op_1454_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_95_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_1454_to_fp16)[name = tensor("tensor_95_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(720317056)))]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(723593920)))]; + tensor linear_91_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_k_proj_weight_to_fp16, x = hidden_states_91_cast_fp16)[name = tensor("linear_91_cast_fp16")]; + tensor var_1459 = const()[name = tensor("op_1459"), val = tensor([1, -1, 20, 64])]; + tensor var_1460_cast_fp16 = reshape(shape = var_1459, x = linear_91_cast_fp16)[name = tensor("op_1460_cast_fp16")]; + tensor var_1461_perm_0 = const()[name = tensor("op_1461_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(723596544)))]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(726873408)))]; + tensor linear_92_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_v_proj_weight_to_fp16, x = hidden_states_91_cast_fp16)[name = tensor("linear_92_cast_fp16")]; + tensor var_1466 = const()[name = tensor("op_1466"), val = tensor([1, -1, 20, 64])]; + tensor var_1467_cast_fp16 = reshape(shape = var_1466, x = linear_92_cast_fp16)[name = tensor("op_1467_cast_fp16")]; + tensor var_1468_perm_0 = const()[name = tensor("op_1468_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1475 = const()[name = tensor("op_1475"), val = tensor([1, 77, 20, 64])]; + tensor var_1476_cast_fp16 = reshape(shape = var_1475, x = tensor_95_cast_fp16)[name = tensor("op_1476_cast_fp16")]; + tensor var_1477_perm_0 = const()[name = tensor("op_1477_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1479 = const()[name = tensor("op_1479"), val = tensor([20, -1, 64])]; + tensor var_1477_cast_fp16 = transpose(perm = var_1477_perm_0, x = var_1476_cast_fp16)[name = tensor("transpose_66")]; + tensor query_states_31_cast_fp16 = reshape(shape = var_1479, x = var_1477_cast_fp16)[name = tensor("query_states_31_cast_fp16")]; + tensor var_1481 = const()[name = tensor("op_1481"), val = tensor([20, -1, 64])]; + tensor var_1461_cast_fp16 = transpose(perm = var_1461_perm_0, x = var_1460_cast_fp16)[name = tensor("transpose_68")]; + tensor key_states_63_cast_fp16 = reshape(shape = var_1481, x = var_1461_cast_fp16)[name = tensor("key_states_63_cast_fp16")]; + tensor var_1483 = const()[name = tensor("op_1483"), val = tensor([20, -1, 64])]; + tensor var_1468_cast_fp16 = transpose(perm = var_1468_perm_0, x = var_1467_cast_fp16)[name = tensor("transpose_67")]; + tensor value_states_63_cast_fp16 = reshape(shape = var_1483, x = var_1468_cast_fp16)[name = tensor("value_states_63_cast_fp16")]; + tensor attn_weights_91_transpose_x_1 = const()[name = tensor("attn_weights_91_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_91_transpose_y_1 = const()[name = tensor("attn_weights_91_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_91_cast_fp16 = matmul(transpose_x = attn_weights_91_transpose_x_1, transpose_y = attn_weights_91_transpose_y_1, x = query_states_31_cast_fp16, y = key_states_63_cast_fp16)[name = tensor("attn_weights_91_cast_fp16")]; + tensor var_1488 = const()[name = tensor("op_1488"), val = tensor([1, 20, 77, 77])]; + tensor var_1489_cast_fp16 = reshape(shape = var_1488, x = attn_weights_91_cast_fp16)[name = tensor("op_1489_cast_fp16")]; + tensor attn_weights_93_cast_fp16 = add(x = var_1489_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_93_cast_fp16")]; + tensor var_1494 = const()[name = tensor("op_1494"), val = tensor([20, 77, 77])]; + tensor input_245_cast_fp16 = reshape(shape = var_1494, x = attn_weights_93_cast_fp16)[name = tensor("input_245_cast_fp16")]; + tensor input_247_cast_fp16 = softmax(axis = var_5, x = input_245_cast_fp16)[name = tensor("input_247_cast_fp16")]; + tensor attn_output_91_transpose_x_0 = const()[name = tensor("attn_output_91_transpose_x_0"), val = tensor(false)]; + tensor attn_output_91_transpose_y_0 = const()[name = tensor("attn_output_91_transpose_y_0"), val = tensor(false)]; + tensor attn_output_91_cast_fp16 = matmul(transpose_x = attn_output_91_transpose_x_0, transpose_y = attn_output_91_transpose_y_0, x = input_247_cast_fp16, y = value_states_63_cast_fp16)[name = tensor("attn_output_91_cast_fp16")]; + tensor var_1499 = const()[name = tensor("op_1499"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_93_cast_fp16 = reshape(shape = var_1499, x = attn_output_91_cast_fp16)[name = tensor("attn_output_93_cast_fp16")]; + tensor attn_output_95_perm_0 = const()[name = tensor("attn_output_95_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1502 = const()[name = tensor("op_1502"), val = tensor([1, 77, 1280])]; + tensor attn_output_95_cast_fp16 = transpose(perm = attn_output_95_perm_0, x = attn_output_93_cast_fp16)[name = tensor("transpose_65")]; + tensor input_249_cast_fp16 = reshape(shape = var_1502, x = attn_output_95_cast_fp16)[name = tensor("input_249_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(726876032)))]; + tensor text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(730152896)))]; + tensor linear_93_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_self_attn_out_proj_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("linear_93_cast_fp16")]; + tensor input_251_cast_fp16 = add(x = input_243_cast_fp16, y = linear_93_cast_fp16)[name = tensor("input_251_cast_fp16")]; + tensor input_253_axes_0 = const()[name = tensor("input_253_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(730155520)))]; + tensor text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(730158144)))]; + tensor input_253_cast_fp16 = layer_norm(axes = input_253_axes_0, beta = text_encoder_text_model_encoder_layers_15_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_15_layer_norm2_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("input_253_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(730160768)))]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(743268032)))]; + tensor linear_94_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_mlp_fc1_weight_to_fp16, x = input_253_cast_fp16)[name = tensor("linear_94_cast_fp16")]; + tensor input_257_mode_0 = const()[name = tensor("input_257_mode_0"), val = tensor("EXACT")]; + tensor input_257_cast_fp16 = gelu(mode = input_257_mode_0, x = linear_94_cast_fp16)[name = tensor("input_257_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(743278336)))]; + tensor text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(756385600)))]; + tensor linear_95_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_15_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_15_mlp_fc2_weight_to_fp16, x = input_257_cast_fp16)[name = tensor("linear_95_cast_fp16")]; + tensor input_259_cast_fp16 = add(x = input_251_cast_fp16, y = linear_95_cast_fp16)[name = tensor("input_259_cast_fp16")]; + tensor hidden_states_97_axes_0 = const()[name = tensor("hidden_states_97_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(756388224)))]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(756390848)))]; + tensor hidden_states_97_cast_fp16 = layer_norm(axes = hidden_states_97_axes_0, beta = text_encoder_text_model_encoder_layers_16_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_16_layer_norm1_weight_to_fp16, x = input_259_cast_fp16)[name = tensor("hidden_states_97_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(756393472)))]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(759670336)))]; + tensor linear_96_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_q_proj_weight_to_fp16, x = hidden_states_97_cast_fp16)[name = tensor("linear_96_cast_fp16")]; + tensor var_1541_to_fp16 = const()[name = tensor("op_1541_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_101_cast_fp16 = mul(x = linear_96_cast_fp16, y = var_1541_to_fp16)[name = tensor("tensor_101_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(759672960)))]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(762949824)))]; + tensor linear_97_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_k_proj_weight_to_fp16, x = hidden_states_97_cast_fp16)[name = tensor("linear_97_cast_fp16")]; + tensor var_1546 = const()[name = tensor("op_1546"), val = tensor([1, -1, 20, 64])]; + tensor var_1547_cast_fp16 = reshape(shape = var_1546, x = linear_97_cast_fp16)[name = tensor("op_1547_cast_fp16")]; + tensor var_1548_perm_0 = const()[name = tensor("op_1548_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(762952448)))]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(766229312)))]; + tensor linear_98_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_v_proj_weight_to_fp16, x = hidden_states_97_cast_fp16)[name = tensor("linear_98_cast_fp16")]; + tensor var_1553 = const()[name = tensor("op_1553"), val = tensor([1, -1, 20, 64])]; + tensor var_1554_cast_fp16 = reshape(shape = var_1553, x = linear_98_cast_fp16)[name = tensor("op_1554_cast_fp16")]; + tensor var_1555_perm_0 = const()[name = tensor("op_1555_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1562 = const()[name = tensor("op_1562"), val = tensor([1, 77, 20, 64])]; + tensor var_1563_cast_fp16 = reshape(shape = var_1562, x = tensor_101_cast_fp16)[name = tensor("op_1563_cast_fp16")]; + tensor var_1564_perm_0 = const()[name = tensor("op_1564_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1566 = const()[name = tensor("op_1566"), val = tensor([20, -1, 64])]; + tensor var_1564_cast_fp16 = transpose(perm = var_1564_perm_0, x = var_1563_cast_fp16)[name = tensor("transpose_62")]; + tensor query_states_33_cast_fp16 = reshape(shape = var_1566, x = var_1564_cast_fp16)[name = tensor("query_states_33_cast_fp16")]; + tensor var_1568 = const()[name = tensor("op_1568"), val = tensor([20, -1, 64])]; + tensor var_1548_cast_fp16 = transpose(perm = var_1548_perm_0, x = var_1547_cast_fp16)[name = tensor("transpose_64")]; + tensor key_states_67_cast_fp16 = reshape(shape = var_1568, x = var_1548_cast_fp16)[name = tensor("key_states_67_cast_fp16")]; + tensor var_1570 = const()[name = tensor("op_1570"), val = tensor([20, -1, 64])]; + tensor var_1555_cast_fp16 = transpose(perm = var_1555_perm_0, x = var_1554_cast_fp16)[name = tensor("transpose_63")]; + tensor value_states_67_cast_fp16 = reshape(shape = var_1570, x = var_1555_cast_fp16)[name = tensor("value_states_67_cast_fp16")]; + tensor attn_weights_97_transpose_x_1 = const()[name = tensor("attn_weights_97_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_97_transpose_y_1 = const()[name = tensor("attn_weights_97_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_97_cast_fp16 = matmul(transpose_x = attn_weights_97_transpose_x_1, transpose_y = attn_weights_97_transpose_y_1, x = query_states_33_cast_fp16, y = key_states_67_cast_fp16)[name = tensor("attn_weights_97_cast_fp16")]; + tensor var_1575 = const()[name = tensor("op_1575"), val = tensor([1, 20, 77, 77])]; + tensor var_1576_cast_fp16 = reshape(shape = var_1575, x = attn_weights_97_cast_fp16)[name = tensor("op_1576_cast_fp16")]; + tensor attn_weights_99_cast_fp16 = add(x = var_1576_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_99_cast_fp16")]; + tensor var_1581 = const()[name = tensor("op_1581"), val = tensor([20, 77, 77])]; + tensor input_261_cast_fp16 = reshape(shape = var_1581, x = attn_weights_99_cast_fp16)[name = tensor("input_261_cast_fp16")]; + tensor input_263_cast_fp16 = softmax(axis = var_5, x = input_261_cast_fp16)[name = tensor("input_263_cast_fp16")]; + tensor attn_output_97_transpose_x_0 = const()[name = tensor("attn_output_97_transpose_x_0"), val = tensor(false)]; + tensor attn_output_97_transpose_y_0 = const()[name = tensor("attn_output_97_transpose_y_0"), val = tensor(false)]; + tensor attn_output_97_cast_fp16 = matmul(transpose_x = attn_output_97_transpose_x_0, transpose_y = attn_output_97_transpose_y_0, x = input_263_cast_fp16, y = value_states_67_cast_fp16)[name = tensor("attn_output_97_cast_fp16")]; + tensor var_1586 = const()[name = tensor("op_1586"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_99_cast_fp16 = reshape(shape = var_1586, x = attn_output_97_cast_fp16)[name = tensor("attn_output_99_cast_fp16")]; + tensor attn_output_101_perm_0 = const()[name = tensor("attn_output_101_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1589 = const()[name = tensor("op_1589"), val = tensor([1, 77, 1280])]; + tensor attn_output_101_cast_fp16 = transpose(perm = attn_output_101_perm_0, x = attn_output_99_cast_fp16)[name = tensor("transpose_61")]; + tensor input_265_cast_fp16 = reshape(shape = var_1589, x = attn_output_101_cast_fp16)[name = tensor("input_265_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(766231936)))]; + tensor text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(769508800)))]; + tensor linear_99_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_self_attn_out_proj_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("linear_99_cast_fp16")]; + tensor input_267_cast_fp16 = add(x = input_259_cast_fp16, y = linear_99_cast_fp16)[name = tensor("input_267_cast_fp16")]; + tensor input_269_axes_0 = const()[name = tensor("input_269_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(769511424)))]; + tensor text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(769514048)))]; + tensor input_269_cast_fp16 = layer_norm(axes = input_269_axes_0, beta = text_encoder_text_model_encoder_layers_16_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_16_layer_norm2_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("input_269_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(769516672)))]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(782623936)))]; + tensor linear_100_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_mlp_fc1_weight_to_fp16, x = input_269_cast_fp16)[name = tensor("linear_100_cast_fp16")]; + tensor input_273_mode_0 = const()[name = tensor("input_273_mode_0"), val = tensor("EXACT")]; + tensor input_273_cast_fp16 = gelu(mode = input_273_mode_0, x = linear_100_cast_fp16)[name = tensor("input_273_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(782634240)))]; + tensor text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(795741504)))]; + tensor linear_101_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_16_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_16_mlp_fc2_weight_to_fp16, x = input_273_cast_fp16)[name = tensor("linear_101_cast_fp16")]; + tensor input_275_cast_fp16 = add(x = input_267_cast_fp16, y = linear_101_cast_fp16)[name = tensor("input_275_cast_fp16")]; + tensor hidden_states_103_axes_0 = const()[name = tensor("hidden_states_103_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(795744128)))]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(795746752)))]; + tensor hidden_states_103_cast_fp16 = layer_norm(axes = hidden_states_103_axes_0, beta = text_encoder_text_model_encoder_layers_17_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_17_layer_norm1_weight_to_fp16, x = input_275_cast_fp16)[name = tensor("hidden_states_103_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(795749376)))]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(799026240)))]; + tensor linear_102_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_q_proj_weight_to_fp16, x = hidden_states_103_cast_fp16)[name = tensor("linear_102_cast_fp16")]; + tensor var_1628_to_fp16 = const()[name = tensor("op_1628_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_107_cast_fp16 = mul(x = linear_102_cast_fp16, y = var_1628_to_fp16)[name = tensor("tensor_107_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(799028864)))]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(802305728)))]; + tensor linear_103_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_k_proj_weight_to_fp16, x = hidden_states_103_cast_fp16)[name = tensor("linear_103_cast_fp16")]; + tensor var_1633 = const()[name = tensor("op_1633"), val = tensor([1, -1, 20, 64])]; + tensor var_1634_cast_fp16 = reshape(shape = var_1633, x = linear_103_cast_fp16)[name = tensor("op_1634_cast_fp16")]; + tensor var_1635_perm_0 = const()[name = tensor("op_1635_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(802308352)))]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(805585216)))]; + tensor linear_104_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_v_proj_weight_to_fp16, x = hidden_states_103_cast_fp16)[name = tensor("linear_104_cast_fp16")]; + tensor var_1640 = const()[name = tensor("op_1640"), val = tensor([1, -1, 20, 64])]; + tensor var_1641_cast_fp16 = reshape(shape = var_1640, x = linear_104_cast_fp16)[name = tensor("op_1641_cast_fp16")]; + tensor var_1642_perm_0 = const()[name = tensor("op_1642_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1649 = const()[name = tensor("op_1649"), val = tensor([1, 77, 20, 64])]; + tensor var_1650_cast_fp16 = reshape(shape = var_1649, x = tensor_107_cast_fp16)[name = tensor("op_1650_cast_fp16")]; + tensor var_1651_perm_0 = const()[name = tensor("op_1651_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1653 = const()[name = tensor("op_1653"), val = tensor([20, -1, 64])]; + tensor var_1651_cast_fp16 = transpose(perm = var_1651_perm_0, x = var_1650_cast_fp16)[name = tensor("transpose_58")]; + tensor query_states_35_cast_fp16 = reshape(shape = var_1653, x = var_1651_cast_fp16)[name = tensor("query_states_35_cast_fp16")]; + tensor var_1655 = const()[name = tensor("op_1655"), val = tensor([20, -1, 64])]; + tensor var_1635_cast_fp16 = transpose(perm = var_1635_perm_0, x = var_1634_cast_fp16)[name = tensor("transpose_60")]; + tensor key_states_71_cast_fp16 = reshape(shape = var_1655, x = var_1635_cast_fp16)[name = tensor("key_states_71_cast_fp16")]; + tensor var_1657 = const()[name = tensor("op_1657"), val = tensor([20, -1, 64])]; + tensor var_1642_cast_fp16 = transpose(perm = var_1642_perm_0, x = var_1641_cast_fp16)[name = tensor("transpose_59")]; + tensor value_states_71_cast_fp16 = reshape(shape = var_1657, x = var_1642_cast_fp16)[name = tensor("value_states_71_cast_fp16")]; + tensor attn_weights_103_transpose_x_1 = const()[name = tensor("attn_weights_103_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_103_transpose_y_1 = const()[name = tensor("attn_weights_103_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_103_cast_fp16 = matmul(transpose_x = attn_weights_103_transpose_x_1, transpose_y = attn_weights_103_transpose_y_1, x = query_states_35_cast_fp16, y = key_states_71_cast_fp16)[name = tensor("attn_weights_103_cast_fp16")]; + tensor var_1662 = const()[name = tensor("op_1662"), val = tensor([1, 20, 77, 77])]; + tensor var_1663_cast_fp16 = reshape(shape = var_1662, x = attn_weights_103_cast_fp16)[name = tensor("op_1663_cast_fp16")]; + tensor attn_weights_105_cast_fp16 = add(x = var_1663_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_105_cast_fp16")]; + tensor var_1668 = const()[name = tensor("op_1668"), val = tensor([20, 77, 77])]; + tensor input_277_cast_fp16 = reshape(shape = var_1668, x = attn_weights_105_cast_fp16)[name = tensor("input_277_cast_fp16")]; + tensor input_279_cast_fp16 = softmax(axis = var_5, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; + tensor attn_output_103_transpose_x_0 = const()[name = tensor("attn_output_103_transpose_x_0"), val = tensor(false)]; + tensor attn_output_103_transpose_y_0 = const()[name = tensor("attn_output_103_transpose_y_0"), val = tensor(false)]; + tensor attn_output_103_cast_fp16 = matmul(transpose_x = attn_output_103_transpose_x_0, transpose_y = attn_output_103_transpose_y_0, x = input_279_cast_fp16, y = value_states_71_cast_fp16)[name = tensor("attn_output_103_cast_fp16")]; + tensor var_1673 = const()[name = tensor("op_1673"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_105_cast_fp16 = reshape(shape = var_1673, x = attn_output_103_cast_fp16)[name = tensor("attn_output_105_cast_fp16")]; + tensor attn_output_107_perm_0 = const()[name = tensor("attn_output_107_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1676 = const()[name = tensor("op_1676"), val = tensor([1, 77, 1280])]; + tensor attn_output_107_cast_fp16 = transpose(perm = attn_output_107_perm_0, x = attn_output_105_cast_fp16)[name = tensor("transpose_57")]; + tensor input_281_cast_fp16 = reshape(shape = var_1676, x = attn_output_107_cast_fp16)[name = tensor("input_281_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(805587840)))]; + tensor text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(808864704)))]; + tensor linear_105_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_self_attn_out_proj_weight_to_fp16, x = input_281_cast_fp16)[name = tensor("linear_105_cast_fp16")]; + tensor input_283_cast_fp16 = add(x = input_275_cast_fp16, y = linear_105_cast_fp16)[name = tensor("input_283_cast_fp16")]; + tensor input_285_axes_0 = const()[name = tensor("input_285_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(808867328)))]; + tensor text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(808869952)))]; + tensor input_285_cast_fp16 = layer_norm(axes = input_285_axes_0, beta = text_encoder_text_model_encoder_layers_17_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_17_layer_norm2_weight_to_fp16, x = input_283_cast_fp16)[name = tensor("input_285_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(808872576)))]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(821979840)))]; + tensor linear_106_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_mlp_fc1_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("linear_106_cast_fp16")]; + tensor input_289_mode_0 = const()[name = tensor("input_289_mode_0"), val = tensor("EXACT")]; + tensor input_289_cast_fp16 = gelu(mode = input_289_mode_0, x = linear_106_cast_fp16)[name = tensor("input_289_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(821990144)))]; + tensor text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(835097408)))]; + tensor linear_107_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_17_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_17_mlp_fc2_weight_to_fp16, x = input_289_cast_fp16)[name = tensor("linear_107_cast_fp16")]; + tensor input_291_cast_fp16 = add(x = input_283_cast_fp16, y = linear_107_cast_fp16)[name = tensor("input_291_cast_fp16")]; + tensor hidden_states_109_axes_0 = const()[name = tensor("hidden_states_109_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(835100032)))]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(835102656)))]; + tensor hidden_states_109_cast_fp16 = layer_norm(axes = hidden_states_109_axes_0, beta = text_encoder_text_model_encoder_layers_18_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_18_layer_norm1_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("hidden_states_109_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(835105280)))]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(838382144)))]; + tensor linear_108_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_q_proj_weight_to_fp16, x = hidden_states_109_cast_fp16)[name = tensor("linear_108_cast_fp16")]; + tensor var_1715_to_fp16 = const()[name = tensor("op_1715_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_113_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_1715_to_fp16)[name = tensor("tensor_113_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(838384768)))]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(841661632)))]; + tensor linear_109_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_k_proj_weight_to_fp16, x = hidden_states_109_cast_fp16)[name = tensor("linear_109_cast_fp16")]; + tensor var_1720 = const()[name = tensor("op_1720"), val = tensor([1, -1, 20, 64])]; + tensor var_1721_cast_fp16 = reshape(shape = var_1720, x = linear_109_cast_fp16)[name = tensor("op_1721_cast_fp16")]; + tensor var_1722_perm_0 = const()[name = tensor("op_1722_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(841664256)))]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(844941120)))]; + tensor linear_110_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_v_proj_weight_to_fp16, x = hidden_states_109_cast_fp16)[name = tensor("linear_110_cast_fp16")]; + tensor var_1727 = const()[name = tensor("op_1727"), val = tensor([1, -1, 20, 64])]; + tensor var_1728_cast_fp16 = reshape(shape = var_1727, x = linear_110_cast_fp16)[name = tensor("op_1728_cast_fp16")]; + tensor var_1729_perm_0 = const()[name = tensor("op_1729_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1736 = const()[name = tensor("op_1736"), val = tensor([1, 77, 20, 64])]; + tensor var_1737_cast_fp16 = reshape(shape = var_1736, x = tensor_113_cast_fp16)[name = tensor("op_1737_cast_fp16")]; + tensor var_1738_perm_0 = const()[name = tensor("op_1738_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1740 = const()[name = tensor("op_1740"), val = tensor([20, -1, 64])]; + tensor var_1738_cast_fp16 = transpose(perm = var_1738_perm_0, x = var_1737_cast_fp16)[name = tensor("transpose_54")]; + tensor query_states_37_cast_fp16 = reshape(shape = var_1740, x = var_1738_cast_fp16)[name = tensor("query_states_37_cast_fp16")]; + tensor var_1742 = const()[name = tensor("op_1742"), val = tensor([20, -1, 64])]; + tensor var_1722_cast_fp16 = transpose(perm = var_1722_perm_0, x = var_1721_cast_fp16)[name = tensor("transpose_56")]; + tensor key_states_75_cast_fp16 = reshape(shape = var_1742, x = var_1722_cast_fp16)[name = tensor("key_states_75_cast_fp16")]; + tensor var_1744 = const()[name = tensor("op_1744"), val = tensor([20, -1, 64])]; + tensor var_1729_cast_fp16 = transpose(perm = var_1729_perm_0, x = var_1728_cast_fp16)[name = tensor("transpose_55")]; + tensor value_states_75_cast_fp16 = reshape(shape = var_1744, x = var_1729_cast_fp16)[name = tensor("value_states_75_cast_fp16")]; + tensor attn_weights_109_transpose_x_1 = const()[name = tensor("attn_weights_109_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_109_transpose_y_1 = const()[name = tensor("attn_weights_109_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_109_cast_fp16 = matmul(transpose_x = attn_weights_109_transpose_x_1, transpose_y = attn_weights_109_transpose_y_1, x = query_states_37_cast_fp16, y = key_states_75_cast_fp16)[name = tensor("attn_weights_109_cast_fp16")]; + tensor var_1749 = const()[name = tensor("op_1749"), val = tensor([1, 20, 77, 77])]; + tensor var_1750_cast_fp16 = reshape(shape = var_1749, x = attn_weights_109_cast_fp16)[name = tensor("op_1750_cast_fp16")]; + tensor attn_weights_111_cast_fp16 = add(x = var_1750_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_111_cast_fp16")]; + tensor var_1755 = const()[name = tensor("op_1755"), val = tensor([20, 77, 77])]; + tensor input_293_cast_fp16 = reshape(shape = var_1755, x = attn_weights_111_cast_fp16)[name = tensor("input_293_cast_fp16")]; + tensor input_295_cast_fp16 = softmax(axis = var_5, x = input_293_cast_fp16)[name = tensor("input_295_cast_fp16")]; + tensor attn_output_109_transpose_x_0 = const()[name = tensor("attn_output_109_transpose_x_0"), val = tensor(false)]; + tensor attn_output_109_transpose_y_0 = const()[name = tensor("attn_output_109_transpose_y_0"), val = tensor(false)]; + tensor attn_output_109_cast_fp16 = matmul(transpose_x = attn_output_109_transpose_x_0, transpose_y = attn_output_109_transpose_y_0, x = input_295_cast_fp16, y = value_states_75_cast_fp16)[name = tensor("attn_output_109_cast_fp16")]; + tensor var_1760 = const()[name = tensor("op_1760"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_111_cast_fp16 = reshape(shape = var_1760, x = attn_output_109_cast_fp16)[name = tensor("attn_output_111_cast_fp16")]; + tensor attn_output_113_perm_0 = const()[name = tensor("attn_output_113_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1763 = const()[name = tensor("op_1763"), val = tensor([1, 77, 1280])]; + tensor attn_output_113_cast_fp16 = transpose(perm = attn_output_113_perm_0, x = attn_output_111_cast_fp16)[name = tensor("transpose_53")]; + tensor input_297_cast_fp16 = reshape(shape = var_1763, x = attn_output_113_cast_fp16)[name = tensor("input_297_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(844943744)))]; + tensor text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(848220608)))]; + tensor linear_111_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_self_attn_out_proj_weight_to_fp16, x = input_297_cast_fp16)[name = tensor("linear_111_cast_fp16")]; + tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_111_cast_fp16)[name = tensor("input_299_cast_fp16")]; + tensor input_301_axes_0 = const()[name = tensor("input_301_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(848223232)))]; + tensor text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(848225856)))]; + tensor input_301_cast_fp16 = layer_norm(axes = input_301_axes_0, beta = text_encoder_text_model_encoder_layers_18_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_18_layer_norm2_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("input_301_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(848228480)))]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(861335744)))]; + tensor linear_112_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_mlp_fc1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("linear_112_cast_fp16")]; + tensor input_305_mode_0 = const()[name = tensor("input_305_mode_0"), val = tensor("EXACT")]; + tensor input_305_cast_fp16 = gelu(mode = input_305_mode_0, x = linear_112_cast_fp16)[name = tensor("input_305_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(861346048)))]; + tensor text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(874453312)))]; + tensor linear_113_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_18_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_18_mlp_fc2_weight_to_fp16, x = input_305_cast_fp16)[name = tensor("linear_113_cast_fp16")]; + tensor input_307_cast_fp16 = add(x = input_299_cast_fp16, y = linear_113_cast_fp16)[name = tensor("input_307_cast_fp16")]; + tensor hidden_states_115_axes_0 = const()[name = tensor("hidden_states_115_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(874455936)))]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(874458560)))]; + tensor hidden_states_115_cast_fp16 = layer_norm(axes = hidden_states_115_axes_0, beta = text_encoder_text_model_encoder_layers_19_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_19_layer_norm1_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("hidden_states_115_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(874461184)))]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(877738048)))]; + tensor linear_114_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_q_proj_weight_to_fp16, x = hidden_states_115_cast_fp16)[name = tensor("linear_114_cast_fp16")]; + tensor var_1802_to_fp16 = const()[name = tensor("op_1802_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_119_cast_fp16 = mul(x = linear_114_cast_fp16, y = var_1802_to_fp16)[name = tensor("tensor_119_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(877740672)))]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(881017536)))]; + tensor linear_115_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_k_proj_weight_to_fp16, x = hidden_states_115_cast_fp16)[name = tensor("linear_115_cast_fp16")]; + tensor var_1807 = const()[name = tensor("op_1807"), val = tensor([1, -1, 20, 64])]; + tensor var_1808_cast_fp16 = reshape(shape = var_1807, x = linear_115_cast_fp16)[name = tensor("op_1808_cast_fp16")]; + tensor var_1809_perm_0 = const()[name = tensor("op_1809_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(881020160)))]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884297024)))]; + tensor linear_116_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_v_proj_weight_to_fp16, x = hidden_states_115_cast_fp16)[name = tensor("linear_116_cast_fp16")]; + tensor var_1814 = const()[name = tensor("op_1814"), val = tensor([1, -1, 20, 64])]; + tensor var_1815_cast_fp16 = reshape(shape = var_1814, x = linear_116_cast_fp16)[name = tensor("op_1815_cast_fp16")]; + tensor var_1816_perm_0 = const()[name = tensor("op_1816_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1823 = const()[name = tensor("op_1823"), val = tensor([1, 77, 20, 64])]; + tensor var_1824_cast_fp16 = reshape(shape = var_1823, x = tensor_119_cast_fp16)[name = tensor("op_1824_cast_fp16")]; + tensor var_1825_perm_0 = const()[name = tensor("op_1825_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1827 = const()[name = tensor("op_1827"), val = tensor([20, -1, 64])]; + tensor var_1825_cast_fp16 = transpose(perm = var_1825_perm_0, x = var_1824_cast_fp16)[name = tensor("transpose_50")]; + tensor query_states_39_cast_fp16 = reshape(shape = var_1827, x = var_1825_cast_fp16)[name = tensor("query_states_39_cast_fp16")]; + tensor var_1829 = const()[name = tensor("op_1829"), val = tensor([20, -1, 64])]; + tensor var_1809_cast_fp16 = transpose(perm = var_1809_perm_0, x = var_1808_cast_fp16)[name = tensor("transpose_52")]; + tensor key_states_79_cast_fp16 = reshape(shape = var_1829, x = var_1809_cast_fp16)[name = tensor("key_states_79_cast_fp16")]; + tensor var_1831 = const()[name = tensor("op_1831"), val = tensor([20, -1, 64])]; + tensor var_1816_cast_fp16 = transpose(perm = var_1816_perm_0, x = var_1815_cast_fp16)[name = tensor("transpose_51")]; + tensor value_states_79_cast_fp16 = reshape(shape = var_1831, x = var_1816_cast_fp16)[name = tensor("value_states_79_cast_fp16")]; + tensor attn_weights_115_transpose_x_1 = const()[name = tensor("attn_weights_115_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_115_transpose_y_1 = const()[name = tensor("attn_weights_115_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_115_cast_fp16 = matmul(transpose_x = attn_weights_115_transpose_x_1, transpose_y = attn_weights_115_transpose_y_1, x = query_states_39_cast_fp16, y = key_states_79_cast_fp16)[name = tensor("attn_weights_115_cast_fp16")]; + tensor var_1836 = const()[name = tensor("op_1836"), val = tensor([1, 20, 77, 77])]; + tensor var_1837_cast_fp16 = reshape(shape = var_1836, x = attn_weights_115_cast_fp16)[name = tensor("op_1837_cast_fp16")]; + tensor attn_weights_117_cast_fp16 = add(x = var_1837_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_117_cast_fp16")]; + tensor var_1842 = const()[name = tensor("op_1842"), val = tensor([20, 77, 77])]; + tensor input_309_cast_fp16 = reshape(shape = var_1842, x = attn_weights_117_cast_fp16)[name = tensor("input_309_cast_fp16")]; + tensor input_311_cast_fp16 = softmax(axis = var_5, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; + tensor attn_output_115_transpose_x_0 = const()[name = tensor("attn_output_115_transpose_x_0"), val = tensor(false)]; + tensor attn_output_115_transpose_y_0 = const()[name = tensor("attn_output_115_transpose_y_0"), val = tensor(false)]; + tensor attn_output_115_cast_fp16 = matmul(transpose_x = attn_output_115_transpose_x_0, transpose_y = attn_output_115_transpose_y_0, x = input_311_cast_fp16, y = value_states_79_cast_fp16)[name = tensor("attn_output_115_cast_fp16")]; + tensor var_1847 = const()[name = tensor("op_1847"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_117_cast_fp16 = reshape(shape = var_1847, x = attn_output_115_cast_fp16)[name = tensor("attn_output_117_cast_fp16")]; + tensor attn_output_119_perm_0 = const()[name = tensor("attn_output_119_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1850 = const()[name = tensor("op_1850"), val = tensor([1, 77, 1280])]; + tensor attn_output_119_cast_fp16 = transpose(perm = attn_output_119_perm_0, x = attn_output_117_cast_fp16)[name = tensor("transpose_49")]; + tensor input_313_cast_fp16 = reshape(shape = var_1850, x = attn_output_119_cast_fp16)[name = tensor("input_313_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(884299648)))]; + tensor text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(887576512)))]; + tensor linear_117_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_self_attn_out_proj_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("linear_117_cast_fp16")]; + tensor input_315_cast_fp16 = add(x = input_307_cast_fp16, y = linear_117_cast_fp16)[name = tensor("input_315_cast_fp16")]; + tensor input_317_axes_0 = const()[name = tensor("input_317_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(887579136)))]; + tensor text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(887581760)))]; + tensor input_317_cast_fp16 = layer_norm(axes = input_317_axes_0, beta = text_encoder_text_model_encoder_layers_19_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_19_layer_norm2_weight_to_fp16, x = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(887584384)))]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(900691648)))]; + tensor linear_118_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_mlp_fc1_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("linear_118_cast_fp16")]; + tensor input_321_mode_0 = const()[name = tensor("input_321_mode_0"), val = tensor("EXACT")]; + tensor input_321_cast_fp16 = gelu(mode = input_321_mode_0, x = linear_118_cast_fp16)[name = tensor("input_321_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(900701952)))]; + tensor text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(913809216)))]; + tensor linear_119_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_19_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_19_mlp_fc2_weight_to_fp16, x = input_321_cast_fp16)[name = tensor("linear_119_cast_fp16")]; + tensor input_323_cast_fp16 = add(x = input_315_cast_fp16, y = linear_119_cast_fp16)[name = tensor("input_323_cast_fp16")]; + tensor hidden_states_121_axes_0 = const()[name = tensor("hidden_states_121_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(913811840)))]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(913814464)))]; + tensor hidden_states_121_cast_fp16 = layer_norm(axes = hidden_states_121_axes_0, beta = text_encoder_text_model_encoder_layers_20_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_20_layer_norm1_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("hidden_states_121_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(913817088)))]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(917093952)))]; + tensor linear_120_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_q_proj_weight_to_fp16, x = hidden_states_121_cast_fp16)[name = tensor("linear_120_cast_fp16")]; + tensor var_1889_to_fp16 = const()[name = tensor("op_1889_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_125_cast_fp16 = mul(x = linear_120_cast_fp16, y = var_1889_to_fp16)[name = tensor("tensor_125_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(917096576)))]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(920373440)))]; + tensor linear_121_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_k_proj_weight_to_fp16, x = hidden_states_121_cast_fp16)[name = tensor("linear_121_cast_fp16")]; + tensor var_1894 = const()[name = tensor("op_1894"), val = tensor([1, -1, 20, 64])]; + tensor var_1895_cast_fp16 = reshape(shape = var_1894, x = linear_121_cast_fp16)[name = tensor("op_1895_cast_fp16")]; + tensor var_1896_perm_0 = const()[name = tensor("op_1896_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(920376064)))]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(923652928)))]; + tensor linear_122_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_v_proj_weight_to_fp16, x = hidden_states_121_cast_fp16)[name = tensor("linear_122_cast_fp16")]; + tensor var_1901 = const()[name = tensor("op_1901"), val = tensor([1, -1, 20, 64])]; + tensor var_1902_cast_fp16 = reshape(shape = var_1901, x = linear_122_cast_fp16)[name = tensor("op_1902_cast_fp16")]; + tensor var_1903_perm_0 = const()[name = tensor("op_1903_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1910 = const()[name = tensor("op_1910"), val = tensor([1, 77, 20, 64])]; + tensor var_1911_cast_fp16 = reshape(shape = var_1910, x = tensor_125_cast_fp16)[name = tensor("op_1911_cast_fp16")]; + tensor var_1912_perm_0 = const()[name = tensor("op_1912_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1914 = const()[name = tensor("op_1914"), val = tensor([20, -1, 64])]; + tensor var_1912_cast_fp16 = transpose(perm = var_1912_perm_0, x = var_1911_cast_fp16)[name = tensor("transpose_46")]; + tensor query_states_41_cast_fp16 = reshape(shape = var_1914, x = var_1912_cast_fp16)[name = tensor("query_states_41_cast_fp16")]; + tensor var_1916 = const()[name = tensor("op_1916"), val = tensor([20, -1, 64])]; + tensor var_1896_cast_fp16 = transpose(perm = var_1896_perm_0, x = var_1895_cast_fp16)[name = tensor("transpose_48")]; + tensor key_states_83_cast_fp16 = reshape(shape = var_1916, x = var_1896_cast_fp16)[name = tensor("key_states_83_cast_fp16")]; + tensor var_1918 = const()[name = tensor("op_1918"), val = tensor([20, -1, 64])]; + tensor var_1903_cast_fp16 = transpose(perm = var_1903_perm_0, x = var_1902_cast_fp16)[name = tensor("transpose_47")]; + tensor value_states_83_cast_fp16 = reshape(shape = var_1918, x = var_1903_cast_fp16)[name = tensor("value_states_83_cast_fp16")]; + tensor attn_weights_121_transpose_x_1 = const()[name = tensor("attn_weights_121_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_121_transpose_y_1 = const()[name = tensor("attn_weights_121_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_121_cast_fp16 = matmul(transpose_x = attn_weights_121_transpose_x_1, transpose_y = attn_weights_121_transpose_y_1, x = query_states_41_cast_fp16, y = key_states_83_cast_fp16)[name = tensor("attn_weights_121_cast_fp16")]; + tensor var_1923 = const()[name = tensor("op_1923"), val = tensor([1, 20, 77, 77])]; + tensor var_1924_cast_fp16 = reshape(shape = var_1923, x = attn_weights_121_cast_fp16)[name = tensor("op_1924_cast_fp16")]; + tensor attn_weights_123_cast_fp16 = add(x = var_1924_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_123_cast_fp16")]; + tensor var_1929 = const()[name = tensor("op_1929"), val = tensor([20, 77, 77])]; + tensor input_325_cast_fp16 = reshape(shape = var_1929, x = attn_weights_123_cast_fp16)[name = tensor("input_325_cast_fp16")]; + tensor input_327_cast_fp16 = softmax(axis = var_5, x = input_325_cast_fp16)[name = tensor("input_327_cast_fp16")]; + tensor attn_output_121_transpose_x_0 = const()[name = tensor("attn_output_121_transpose_x_0"), val = tensor(false)]; + tensor attn_output_121_transpose_y_0 = const()[name = tensor("attn_output_121_transpose_y_0"), val = tensor(false)]; + tensor attn_output_121_cast_fp16 = matmul(transpose_x = attn_output_121_transpose_x_0, transpose_y = attn_output_121_transpose_y_0, x = input_327_cast_fp16, y = value_states_83_cast_fp16)[name = tensor("attn_output_121_cast_fp16")]; + tensor var_1934 = const()[name = tensor("op_1934"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_123_cast_fp16 = reshape(shape = var_1934, x = attn_output_121_cast_fp16)[name = tensor("attn_output_123_cast_fp16")]; + tensor attn_output_125_perm_0 = const()[name = tensor("attn_output_125_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1937 = const()[name = tensor("op_1937"), val = tensor([1, 77, 1280])]; + tensor attn_output_125_cast_fp16 = transpose(perm = attn_output_125_perm_0, x = attn_output_123_cast_fp16)[name = tensor("transpose_45")]; + tensor input_329_cast_fp16 = reshape(shape = var_1937, x = attn_output_125_cast_fp16)[name = tensor("input_329_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(923655552)))]; + tensor text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(926932416)))]; + tensor linear_123_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_self_attn_out_proj_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("linear_123_cast_fp16")]; + tensor input_331_cast_fp16 = add(x = input_323_cast_fp16, y = linear_123_cast_fp16)[name = tensor("input_331_cast_fp16")]; + tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(926935040)))]; + tensor text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(926937664)))]; + tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = text_encoder_text_model_encoder_layers_20_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_20_layer_norm2_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(926940288)))]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(940047552)))]; + tensor linear_124_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_mlp_fc1_weight_to_fp16, x = input_333_cast_fp16)[name = tensor("linear_124_cast_fp16")]; + tensor input_337_mode_0 = const()[name = tensor("input_337_mode_0"), val = tensor("EXACT")]; + tensor input_337_cast_fp16 = gelu(mode = input_337_mode_0, x = linear_124_cast_fp16)[name = tensor("input_337_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(940057856)))]; + tensor text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(953165120)))]; + tensor linear_125_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_20_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_20_mlp_fc2_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("linear_125_cast_fp16")]; + tensor input_339_cast_fp16 = add(x = input_331_cast_fp16, y = linear_125_cast_fp16)[name = tensor("input_339_cast_fp16")]; + tensor hidden_states_127_axes_0 = const()[name = tensor("hidden_states_127_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(953167744)))]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(953170368)))]; + tensor hidden_states_127_cast_fp16 = layer_norm(axes = hidden_states_127_axes_0, beta = text_encoder_text_model_encoder_layers_21_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_21_layer_norm1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("hidden_states_127_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(953172992)))]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(956449856)))]; + tensor linear_126_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_q_proj_weight_to_fp16, x = hidden_states_127_cast_fp16)[name = tensor("linear_126_cast_fp16")]; + tensor var_1976_to_fp16 = const()[name = tensor("op_1976_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_131_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_1976_to_fp16)[name = tensor("tensor_131_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(956452480)))]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(959729344)))]; + tensor linear_127_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_k_proj_weight_to_fp16, x = hidden_states_127_cast_fp16)[name = tensor("linear_127_cast_fp16")]; + tensor var_1981 = const()[name = tensor("op_1981"), val = tensor([1, -1, 20, 64])]; + tensor var_1982_cast_fp16 = reshape(shape = var_1981, x = linear_127_cast_fp16)[name = tensor("op_1982_cast_fp16")]; + tensor var_1983_perm_0 = const()[name = tensor("op_1983_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(959731968)))]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(963008832)))]; + tensor linear_128_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_v_proj_weight_to_fp16, x = hidden_states_127_cast_fp16)[name = tensor("linear_128_cast_fp16")]; + tensor var_1988 = const()[name = tensor("op_1988"), val = tensor([1, -1, 20, 64])]; + tensor var_1989_cast_fp16 = reshape(shape = var_1988, x = linear_128_cast_fp16)[name = tensor("op_1989_cast_fp16")]; + tensor var_1990_perm_0 = const()[name = tensor("op_1990_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_1997 = const()[name = tensor("op_1997"), val = tensor([1, 77, 20, 64])]; + tensor var_1998_cast_fp16 = reshape(shape = var_1997, x = tensor_131_cast_fp16)[name = tensor("op_1998_cast_fp16")]; + tensor var_1999_perm_0 = const()[name = tensor("op_1999_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2001 = const()[name = tensor("op_2001"), val = tensor([20, -1, 64])]; + tensor var_1999_cast_fp16 = transpose(perm = var_1999_perm_0, x = var_1998_cast_fp16)[name = tensor("transpose_42")]; + tensor query_states_43_cast_fp16 = reshape(shape = var_2001, x = var_1999_cast_fp16)[name = tensor("query_states_43_cast_fp16")]; + tensor var_2003 = const()[name = tensor("op_2003"), val = tensor([20, -1, 64])]; + tensor var_1983_cast_fp16 = transpose(perm = var_1983_perm_0, x = var_1982_cast_fp16)[name = tensor("transpose_44")]; + tensor key_states_87_cast_fp16 = reshape(shape = var_2003, x = var_1983_cast_fp16)[name = tensor("key_states_87_cast_fp16")]; + tensor var_2005 = const()[name = tensor("op_2005"), val = tensor([20, -1, 64])]; + tensor var_1990_cast_fp16 = transpose(perm = var_1990_perm_0, x = var_1989_cast_fp16)[name = tensor("transpose_43")]; + tensor value_states_87_cast_fp16 = reshape(shape = var_2005, x = var_1990_cast_fp16)[name = tensor("value_states_87_cast_fp16")]; + tensor attn_weights_127_transpose_x_1 = const()[name = tensor("attn_weights_127_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_127_transpose_y_1 = const()[name = tensor("attn_weights_127_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_127_cast_fp16 = matmul(transpose_x = attn_weights_127_transpose_x_1, transpose_y = attn_weights_127_transpose_y_1, x = query_states_43_cast_fp16, y = key_states_87_cast_fp16)[name = tensor("attn_weights_127_cast_fp16")]; + tensor var_2010 = const()[name = tensor("op_2010"), val = tensor([1, 20, 77, 77])]; + tensor var_2011_cast_fp16 = reshape(shape = var_2010, x = attn_weights_127_cast_fp16)[name = tensor("op_2011_cast_fp16")]; + tensor attn_weights_129_cast_fp16 = add(x = var_2011_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_129_cast_fp16")]; + tensor var_2016 = const()[name = tensor("op_2016"), val = tensor([20, 77, 77])]; + tensor input_341_cast_fp16 = reshape(shape = var_2016, x = attn_weights_129_cast_fp16)[name = tensor("input_341_cast_fp16")]; + tensor input_343_cast_fp16 = softmax(axis = var_5, x = input_341_cast_fp16)[name = tensor("input_343_cast_fp16")]; + tensor attn_output_127_transpose_x_0 = const()[name = tensor("attn_output_127_transpose_x_0"), val = tensor(false)]; + tensor attn_output_127_transpose_y_0 = const()[name = tensor("attn_output_127_transpose_y_0"), val = tensor(false)]; + tensor attn_output_127_cast_fp16 = matmul(transpose_x = attn_output_127_transpose_x_0, transpose_y = attn_output_127_transpose_y_0, x = input_343_cast_fp16, y = value_states_87_cast_fp16)[name = tensor("attn_output_127_cast_fp16")]; + tensor var_2021 = const()[name = tensor("op_2021"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_129_cast_fp16 = reshape(shape = var_2021, x = attn_output_127_cast_fp16)[name = tensor("attn_output_129_cast_fp16")]; + tensor attn_output_131_perm_0 = const()[name = tensor("attn_output_131_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2024 = const()[name = tensor("op_2024"), val = tensor([1, 77, 1280])]; + tensor attn_output_131_cast_fp16 = transpose(perm = attn_output_131_perm_0, x = attn_output_129_cast_fp16)[name = tensor("transpose_41")]; + tensor input_345_cast_fp16 = reshape(shape = var_2024, x = attn_output_131_cast_fp16)[name = tensor("input_345_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(963011456)))]; + tensor text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(966288320)))]; + tensor linear_129_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_self_attn_out_proj_weight_to_fp16, x = input_345_cast_fp16)[name = tensor("linear_129_cast_fp16")]; + tensor input_347_cast_fp16 = add(x = input_339_cast_fp16, y = linear_129_cast_fp16)[name = tensor("input_347_cast_fp16")]; + tensor input_349_axes_0 = const()[name = tensor("input_349_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(966290944)))]; + tensor text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(966293568)))]; + tensor input_349_cast_fp16 = layer_norm(axes = input_349_axes_0, beta = text_encoder_text_model_encoder_layers_21_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_21_layer_norm2_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("input_349_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(966296192)))]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(979403456)))]; + tensor linear_130_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_mlp_fc1_weight_to_fp16, x = input_349_cast_fp16)[name = tensor("linear_130_cast_fp16")]; + tensor input_353_mode_0 = const()[name = tensor("input_353_mode_0"), val = tensor("EXACT")]; + tensor input_353_cast_fp16 = gelu(mode = input_353_mode_0, x = linear_130_cast_fp16)[name = tensor("input_353_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(979413760)))]; + tensor text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(992521024)))]; + tensor linear_131_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_21_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_21_mlp_fc2_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("linear_131_cast_fp16")]; + tensor input_355_cast_fp16 = add(x = input_347_cast_fp16, y = linear_131_cast_fp16)[name = tensor("input_355_cast_fp16")]; + tensor hidden_states_133_axes_0 = const()[name = tensor("hidden_states_133_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(992523648)))]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(992526272)))]; + tensor hidden_states_133_cast_fp16 = layer_norm(axes = hidden_states_133_axes_0, beta = text_encoder_text_model_encoder_layers_22_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_22_layer_norm1_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("hidden_states_133_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(992528896)))]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(995805760)))]; + tensor linear_132_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_q_proj_weight_to_fp16, x = hidden_states_133_cast_fp16)[name = tensor("linear_132_cast_fp16")]; + tensor var_2063_to_fp16 = const()[name = tensor("op_2063_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_137_cast_fp16 = mul(x = linear_132_cast_fp16, y = var_2063_to_fp16)[name = tensor("tensor_137_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(995808384)))]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(999085248)))]; + tensor linear_133_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_k_proj_weight_to_fp16, x = hidden_states_133_cast_fp16)[name = tensor("linear_133_cast_fp16")]; + tensor var_2068 = const()[name = tensor("op_2068"), val = tensor([1, -1, 20, 64])]; + tensor var_2069_cast_fp16 = reshape(shape = var_2068, x = linear_133_cast_fp16)[name = tensor("op_2069_cast_fp16")]; + tensor var_2070_perm_0 = const()[name = tensor("op_2070_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(999087872)))]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1002364736)))]; + tensor linear_134_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_v_proj_weight_to_fp16, x = hidden_states_133_cast_fp16)[name = tensor("linear_134_cast_fp16")]; + tensor var_2075 = const()[name = tensor("op_2075"), val = tensor([1, -1, 20, 64])]; + tensor var_2076_cast_fp16 = reshape(shape = var_2075, x = linear_134_cast_fp16)[name = tensor("op_2076_cast_fp16")]; + tensor var_2077_perm_0 = const()[name = tensor("op_2077_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2084 = const()[name = tensor("op_2084"), val = tensor([1, 77, 20, 64])]; + tensor var_2085_cast_fp16 = reshape(shape = var_2084, x = tensor_137_cast_fp16)[name = tensor("op_2085_cast_fp16")]; + tensor var_2086_perm_0 = const()[name = tensor("op_2086_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2088 = const()[name = tensor("op_2088"), val = tensor([20, -1, 64])]; + tensor var_2086_cast_fp16 = transpose(perm = var_2086_perm_0, x = var_2085_cast_fp16)[name = tensor("transpose_38")]; + tensor query_states_45_cast_fp16 = reshape(shape = var_2088, x = var_2086_cast_fp16)[name = tensor("query_states_45_cast_fp16")]; + tensor var_2090 = const()[name = tensor("op_2090"), val = tensor([20, -1, 64])]; + tensor var_2070_cast_fp16 = transpose(perm = var_2070_perm_0, x = var_2069_cast_fp16)[name = tensor("transpose_40")]; + tensor key_states_91_cast_fp16 = reshape(shape = var_2090, x = var_2070_cast_fp16)[name = tensor("key_states_91_cast_fp16")]; + tensor var_2092 = const()[name = tensor("op_2092"), val = tensor([20, -1, 64])]; + tensor var_2077_cast_fp16 = transpose(perm = var_2077_perm_0, x = var_2076_cast_fp16)[name = tensor("transpose_39")]; + tensor value_states_91_cast_fp16 = reshape(shape = var_2092, x = var_2077_cast_fp16)[name = tensor("value_states_91_cast_fp16")]; + tensor attn_weights_133_transpose_x_1 = const()[name = tensor("attn_weights_133_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_133_transpose_y_1 = const()[name = tensor("attn_weights_133_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_133_cast_fp16 = matmul(transpose_x = attn_weights_133_transpose_x_1, transpose_y = attn_weights_133_transpose_y_1, x = query_states_45_cast_fp16, y = key_states_91_cast_fp16)[name = tensor("attn_weights_133_cast_fp16")]; + tensor var_2097 = const()[name = tensor("op_2097"), val = tensor([1, 20, 77, 77])]; + tensor var_2098_cast_fp16 = reshape(shape = var_2097, x = attn_weights_133_cast_fp16)[name = tensor("op_2098_cast_fp16")]; + tensor attn_weights_135_cast_fp16 = add(x = var_2098_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_135_cast_fp16")]; + tensor var_2103 = const()[name = tensor("op_2103"), val = tensor([20, 77, 77])]; + tensor input_357_cast_fp16 = reshape(shape = var_2103, x = attn_weights_135_cast_fp16)[name = tensor("input_357_cast_fp16")]; + tensor input_359_cast_fp16 = softmax(axis = var_5, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; + tensor attn_output_133_transpose_x_0 = const()[name = tensor("attn_output_133_transpose_x_0"), val = tensor(false)]; + tensor attn_output_133_transpose_y_0 = const()[name = tensor("attn_output_133_transpose_y_0"), val = tensor(false)]; + tensor attn_output_133_cast_fp16 = matmul(transpose_x = attn_output_133_transpose_x_0, transpose_y = attn_output_133_transpose_y_0, x = input_359_cast_fp16, y = value_states_91_cast_fp16)[name = tensor("attn_output_133_cast_fp16")]; + tensor var_2108 = const()[name = tensor("op_2108"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_135_cast_fp16 = reshape(shape = var_2108, x = attn_output_133_cast_fp16)[name = tensor("attn_output_135_cast_fp16")]; + tensor attn_output_137_perm_0 = const()[name = tensor("attn_output_137_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2111 = const()[name = tensor("op_2111"), val = tensor([1, 77, 1280])]; + tensor attn_output_137_cast_fp16 = transpose(perm = attn_output_137_perm_0, x = attn_output_135_cast_fp16)[name = tensor("transpose_37")]; + tensor input_361_cast_fp16 = reshape(shape = var_2111, x = attn_output_137_cast_fp16)[name = tensor("input_361_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1002367360)))]; + tensor text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1005644224)))]; + tensor linear_135_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_self_attn_out_proj_weight_to_fp16, x = input_361_cast_fp16)[name = tensor("linear_135_cast_fp16")]; + tensor input_363_cast_fp16 = add(x = input_355_cast_fp16, y = linear_135_cast_fp16)[name = tensor("input_363_cast_fp16")]; + tensor input_365_axes_0 = const()[name = tensor("input_365_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1005646848)))]; + tensor text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1005649472)))]; + tensor input_365_cast_fp16 = layer_norm(axes = input_365_axes_0, beta = text_encoder_text_model_encoder_layers_22_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_22_layer_norm2_weight_to_fp16, x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1005652096)))]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1018759360)))]; + tensor linear_136_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_mlp_fc1_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("linear_136_cast_fp16")]; + tensor input_369_mode_0 = const()[name = tensor("input_369_mode_0"), val = tensor("EXACT")]; + tensor input_369_cast_fp16 = gelu(mode = input_369_mode_0, x = linear_136_cast_fp16)[name = tensor("input_369_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1018769664)))]; + tensor text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1031876928)))]; + tensor linear_137_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_22_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_22_mlp_fc2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("linear_137_cast_fp16")]; + tensor input_371_cast_fp16 = add(x = input_363_cast_fp16, y = linear_137_cast_fp16)[name = tensor("input_371_cast_fp16")]; + tensor hidden_states_139_axes_0 = const()[name = tensor("hidden_states_139_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1031879552)))]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1031882176)))]; + tensor hidden_states_139_cast_fp16 = layer_norm(axes = hidden_states_139_axes_0, beta = text_encoder_text_model_encoder_layers_23_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_23_layer_norm1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("hidden_states_139_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1031884800)))]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1035161664)))]; + tensor linear_138_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_q_proj_weight_to_fp16, x = hidden_states_139_cast_fp16)[name = tensor("linear_138_cast_fp16")]; + tensor var_2150_to_fp16 = const()[name = tensor("op_2150_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_143_cast_fp16 = mul(x = linear_138_cast_fp16, y = var_2150_to_fp16)[name = tensor("tensor_143_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1035164288)))]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1038441152)))]; + tensor linear_139_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_k_proj_weight_to_fp16, x = hidden_states_139_cast_fp16)[name = tensor("linear_139_cast_fp16")]; + tensor var_2155 = const()[name = tensor("op_2155"), val = tensor([1, -1, 20, 64])]; + tensor var_2156_cast_fp16 = reshape(shape = var_2155, x = linear_139_cast_fp16)[name = tensor("op_2156_cast_fp16")]; + tensor var_2157_perm_0 = const()[name = tensor("op_2157_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1038443776)))]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1041720640)))]; + tensor linear_140_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_v_proj_weight_to_fp16, x = hidden_states_139_cast_fp16)[name = tensor("linear_140_cast_fp16")]; + tensor var_2162 = const()[name = tensor("op_2162"), val = tensor([1, -1, 20, 64])]; + tensor var_2163_cast_fp16 = reshape(shape = var_2162, x = linear_140_cast_fp16)[name = tensor("op_2163_cast_fp16")]; + tensor var_2164_perm_0 = const()[name = tensor("op_2164_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2171 = const()[name = tensor("op_2171"), val = tensor([1, 77, 20, 64])]; + tensor var_2172_cast_fp16 = reshape(shape = var_2171, x = tensor_143_cast_fp16)[name = tensor("op_2172_cast_fp16")]; + tensor var_2173_perm_0 = const()[name = tensor("op_2173_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2175 = const()[name = tensor("op_2175"), val = tensor([20, -1, 64])]; + tensor var_2173_cast_fp16 = transpose(perm = var_2173_perm_0, x = var_2172_cast_fp16)[name = tensor("transpose_34")]; + tensor query_states_47_cast_fp16 = reshape(shape = var_2175, x = var_2173_cast_fp16)[name = tensor("query_states_47_cast_fp16")]; + tensor var_2177 = const()[name = tensor("op_2177"), val = tensor([20, -1, 64])]; + tensor var_2157_cast_fp16 = transpose(perm = var_2157_perm_0, x = var_2156_cast_fp16)[name = tensor("transpose_36")]; + tensor key_states_95_cast_fp16 = reshape(shape = var_2177, x = var_2157_cast_fp16)[name = tensor("key_states_95_cast_fp16")]; + tensor var_2179 = const()[name = tensor("op_2179"), val = tensor([20, -1, 64])]; + tensor var_2164_cast_fp16 = transpose(perm = var_2164_perm_0, x = var_2163_cast_fp16)[name = tensor("transpose_35")]; + tensor value_states_95_cast_fp16 = reshape(shape = var_2179, x = var_2164_cast_fp16)[name = tensor("value_states_95_cast_fp16")]; + tensor attn_weights_139_transpose_x_1 = const()[name = tensor("attn_weights_139_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_139_transpose_y_1 = const()[name = tensor("attn_weights_139_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_139_cast_fp16 = matmul(transpose_x = attn_weights_139_transpose_x_1, transpose_y = attn_weights_139_transpose_y_1, x = query_states_47_cast_fp16, y = key_states_95_cast_fp16)[name = tensor("attn_weights_139_cast_fp16")]; + tensor var_2184 = const()[name = tensor("op_2184"), val = tensor([1, 20, 77, 77])]; + tensor var_2185_cast_fp16 = reshape(shape = var_2184, x = attn_weights_139_cast_fp16)[name = tensor("op_2185_cast_fp16")]; + tensor attn_weights_141_cast_fp16 = add(x = var_2185_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_141_cast_fp16")]; + tensor var_2190 = const()[name = tensor("op_2190"), val = tensor([20, 77, 77])]; + tensor input_373_cast_fp16 = reshape(shape = var_2190, x = attn_weights_141_cast_fp16)[name = tensor("input_373_cast_fp16")]; + tensor input_375_cast_fp16 = softmax(axis = var_5, x = input_373_cast_fp16)[name = tensor("input_375_cast_fp16")]; + tensor attn_output_139_transpose_x_0 = const()[name = tensor("attn_output_139_transpose_x_0"), val = tensor(false)]; + tensor attn_output_139_transpose_y_0 = const()[name = tensor("attn_output_139_transpose_y_0"), val = tensor(false)]; + tensor attn_output_139_cast_fp16 = matmul(transpose_x = attn_output_139_transpose_x_0, transpose_y = attn_output_139_transpose_y_0, x = input_375_cast_fp16, y = value_states_95_cast_fp16)[name = tensor("attn_output_139_cast_fp16")]; + tensor var_2195 = const()[name = tensor("op_2195"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_141_cast_fp16 = reshape(shape = var_2195, x = attn_output_139_cast_fp16)[name = tensor("attn_output_141_cast_fp16")]; + tensor attn_output_143_perm_0 = const()[name = tensor("attn_output_143_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2198 = const()[name = tensor("op_2198"), val = tensor([1, 77, 1280])]; + tensor attn_output_143_cast_fp16 = transpose(perm = attn_output_143_perm_0, x = attn_output_141_cast_fp16)[name = tensor("transpose_33")]; + tensor input_377_cast_fp16 = reshape(shape = var_2198, x = attn_output_143_cast_fp16)[name = tensor("input_377_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1041723264)))]; + tensor text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1045000128)))]; + tensor linear_141_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_self_attn_out_proj_weight_to_fp16, x = input_377_cast_fp16)[name = tensor("linear_141_cast_fp16")]; + tensor input_379_cast_fp16 = add(x = input_371_cast_fp16, y = linear_141_cast_fp16)[name = tensor("input_379_cast_fp16")]; + tensor input_381_axes_0 = const()[name = tensor("input_381_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1045002752)))]; + tensor text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1045005376)))]; + tensor input_381_cast_fp16 = layer_norm(axes = input_381_axes_0, beta = text_encoder_text_model_encoder_layers_23_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_23_layer_norm2_weight_to_fp16, x = input_379_cast_fp16)[name = tensor("input_381_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1045008000)))]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1058115264)))]; + tensor linear_142_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_mlp_fc1_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("linear_142_cast_fp16")]; + tensor input_385_mode_0 = const()[name = tensor("input_385_mode_0"), val = tensor("EXACT")]; + tensor input_385_cast_fp16 = gelu(mode = input_385_mode_0, x = linear_142_cast_fp16)[name = tensor("input_385_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1058125568)))]; + tensor text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1071232832)))]; + tensor linear_143_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_23_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_23_mlp_fc2_weight_to_fp16, x = input_385_cast_fp16)[name = tensor("linear_143_cast_fp16")]; + tensor input_387_cast_fp16 = add(x = input_379_cast_fp16, y = linear_143_cast_fp16)[name = tensor("input_387_cast_fp16")]; + tensor hidden_states_145_axes_0 = const()[name = tensor("hidden_states_145_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1071235456)))]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1071238080)))]; + tensor hidden_states_145_cast_fp16 = layer_norm(axes = hidden_states_145_axes_0, beta = text_encoder_text_model_encoder_layers_24_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_24_layer_norm1_weight_to_fp16, x = input_387_cast_fp16)[name = tensor("hidden_states_145_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1071240704)))]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1074517568)))]; + tensor linear_144_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_q_proj_weight_to_fp16, x = hidden_states_145_cast_fp16)[name = tensor("linear_144_cast_fp16")]; + tensor var_2237_to_fp16 = const()[name = tensor("op_2237_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_149_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_2237_to_fp16)[name = tensor("tensor_149_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1074520192)))]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077797056)))]; + tensor linear_145_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_k_proj_weight_to_fp16, x = hidden_states_145_cast_fp16)[name = tensor("linear_145_cast_fp16")]; + tensor var_2242 = const()[name = tensor("op_2242"), val = tensor([1, -1, 20, 64])]; + tensor var_2243_cast_fp16 = reshape(shape = var_2242, x = linear_145_cast_fp16)[name = tensor("op_2243_cast_fp16")]; + tensor var_2244_perm_0 = const()[name = tensor("op_2244_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1077799680)))]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1081076544)))]; + tensor linear_146_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_v_proj_weight_to_fp16, x = hidden_states_145_cast_fp16)[name = tensor("linear_146_cast_fp16")]; + tensor var_2249 = const()[name = tensor("op_2249"), val = tensor([1, -1, 20, 64])]; + tensor var_2250_cast_fp16 = reshape(shape = var_2249, x = linear_146_cast_fp16)[name = tensor("op_2250_cast_fp16")]; + tensor var_2251_perm_0 = const()[name = tensor("op_2251_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2258 = const()[name = tensor("op_2258"), val = tensor([1, 77, 20, 64])]; + tensor var_2259_cast_fp16 = reshape(shape = var_2258, x = tensor_149_cast_fp16)[name = tensor("op_2259_cast_fp16")]; + tensor var_2260_perm_0 = const()[name = tensor("op_2260_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2262 = const()[name = tensor("op_2262"), val = tensor([20, -1, 64])]; + tensor var_2260_cast_fp16 = transpose(perm = var_2260_perm_0, x = var_2259_cast_fp16)[name = tensor("transpose_30")]; + tensor query_states_49_cast_fp16 = reshape(shape = var_2262, x = var_2260_cast_fp16)[name = tensor("query_states_49_cast_fp16")]; + tensor var_2264 = const()[name = tensor("op_2264"), val = tensor([20, -1, 64])]; + tensor var_2244_cast_fp16 = transpose(perm = var_2244_perm_0, x = var_2243_cast_fp16)[name = tensor("transpose_32")]; + tensor key_states_99_cast_fp16 = reshape(shape = var_2264, x = var_2244_cast_fp16)[name = tensor("key_states_99_cast_fp16")]; + tensor var_2266 = const()[name = tensor("op_2266"), val = tensor([20, -1, 64])]; + tensor var_2251_cast_fp16 = transpose(perm = var_2251_perm_0, x = var_2250_cast_fp16)[name = tensor("transpose_31")]; + tensor value_states_99_cast_fp16 = reshape(shape = var_2266, x = var_2251_cast_fp16)[name = tensor("value_states_99_cast_fp16")]; + tensor attn_weights_145_transpose_x_1 = const()[name = tensor("attn_weights_145_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_145_transpose_y_1 = const()[name = tensor("attn_weights_145_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_145_cast_fp16 = matmul(transpose_x = attn_weights_145_transpose_x_1, transpose_y = attn_weights_145_transpose_y_1, x = query_states_49_cast_fp16, y = key_states_99_cast_fp16)[name = tensor("attn_weights_145_cast_fp16")]; + tensor var_2271 = const()[name = tensor("op_2271"), val = tensor([1, 20, 77, 77])]; + tensor var_2272_cast_fp16 = reshape(shape = var_2271, x = attn_weights_145_cast_fp16)[name = tensor("op_2272_cast_fp16")]; + tensor attn_weights_147_cast_fp16 = add(x = var_2272_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_147_cast_fp16")]; + tensor var_2277 = const()[name = tensor("op_2277"), val = tensor([20, 77, 77])]; + tensor input_389_cast_fp16 = reshape(shape = var_2277, x = attn_weights_147_cast_fp16)[name = tensor("input_389_cast_fp16")]; + tensor input_391_cast_fp16 = softmax(axis = var_5, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; + tensor attn_output_145_transpose_x_0 = const()[name = tensor("attn_output_145_transpose_x_0"), val = tensor(false)]; + tensor attn_output_145_transpose_y_0 = const()[name = tensor("attn_output_145_transpose_y_0"), val = tensor(false)]; + tensor attn_output_145_cast_fp16 = matmul(transpose_x = attn_output_145_transpose_x_0, transpose_y = attn_output_145_transpose_y_0, x = input_391_cast_fp16, y = value_states_99_cast_fp16)[name = tensor("attn_output_145_cast_fp16")]; + tensor var_2282 = const()[name = tensor("op_2282"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_147_cast_fp16 = reshape(shape = var_2282, x = attn_output_145_cast_fp16)[name = tensor("attn_output_147_cast_fp16")]; + tensor attn_output_149_perm_0 = const()[name = tensor("attn_output_149_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2285 = const()[name = tensor("op_2285"), val = tensor([1, 77, 1280])]; + tensor attn_output_149_cast_fp16 = transpose(perm = attn_output_149_perm_0, x = attn_output_147_cast_fp16)[name = tensor("transpose_29")]; + tensor input_393_cast_fp16 = reshape(shape = var_2285, x = attn_output_149_cast_fp16)[name = tensor("input_393_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1081079168)))]; + tensor text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1084356032)))]; + tensor linear_147_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_self_attn_out_proj_weight_to_fp16, x = input_393_cast_fp16)[name = tensor("linear_147_cast_fp16")]; + tensor input_395_cast_fp16 = add(x = input_387_cast_fp16, y = linear_147_cast_fp16)[name = tensor("input_395_cast_fp16")]; + tensor input_397_axes_0 = const()[name = tensor("input_397_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1084358656)))]; + tensor text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1084361280)))]; + tensor input_397_cast_fp16 = layer_norm(axes = input_397_axes_0, beta = text_encoder_text_model_encoder_layers_24_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_24_layer_norm2_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("input_397_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1084363904)))]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1097471168)))]; + tensor linear_148_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_mlp_fc1_weight_to_fp16, x = input_397_cast_fp16)[name = tensor("linear_148_cast_fp16")]; + tensor input_401_mode_0 = const()[name = tensor("input_401_mode_0"), val = tensor("EXACT")]; + tensor input_401_cast_fp16 = gelu(mode = input_401_mode_0, x = linear_148_cast_fp16)[name = tensor("input_401_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1097481472)))]; + tensor text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1110588736)))]; + tensor linear_149_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_24_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_24_mlp_fc2_weight_to_fp16, x = input_401_cast_fp16)[name = tensor("linear_149_cast_fp16")]; + tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_149_cast_fp16)[name = tensor("input_403_cast_fp16")]; + tensor hidden_states_151_axes_0 = const()[name = tensor("hidden_states_151_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1110591360)))]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1110593984)))]; + tensor hidden_states_151_cast_fp16 = layer_norm(axes = hidden_states_151_axes_0, beta = text_encoder_text_model_encoder_layers_25_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_25_layer_norm1_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("hidden_states_151_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1110596608)))]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1113873472)))]; + tensor linear_150_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_q_proj_weight_to_fp16, x = hidden_states_151_cast_fp16)[name = tensor("linear_150_cast_fp16")]; + tensor var_2324_to_fp16 = const()[name = tensor("op_2324_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_155_cast_fp16 = mul(x = linear_150_cast_fp16, y = var_2324_to_fp16)[name = tensor("tensor_155_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1113876096)))]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1117152960)))]; + tensor linear_151_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_k_proj_weight_to_fp16, x = hidden_states_151_cast_fp16)[name = tensor("linear_151_cast_fp16")]; + tensor var_2329 = const()[name = tensor("op_2329"), val = tensor([1, -1, 20, 64])]; + tensor var_2330_cast_fp16 = reshape(shape = var_2329, x = linear_151_cast_fp16)[name = tensor("op_2330_cast_fp16")]; + tensor var_2331_perm_0 = const()[name = tensor("op_2331_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1117155584)))]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1120432448)))]; + tensor linear_152_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_v_proj_weight_to_fp16, x = hidden_states_151_cast_fp16)[name = tensor("linear_152_cast_fp16")]; + tensor var_2336 = const()[name = tensor("op_2336"), val = tensor([1, -1, 20, 64])]; + tensor var_2337_cast_fp16 = reshape(shape = var_2336, x = linear_152_cast_fp16)[name = tensor("op_2337_cast_fp16")]; + tensor var_2338_perm_0 = const()[name = tensor("op_2338_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2345 = const()[name = tensor("op_2345"), val = tensor([1, 77, 20, 64])]; + tensor var_2346_cast_fp16 = reshape(shape = var_2345, x = tensor_155_cast_fp16)[name = tensor("op_2346_cast_fp16")]; + tensor var_2347_perm_0 = const()[name = tensor("op_2347_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2349 = const()[name = tensor("op_2349"), val = tensor([20, -1, 64])]; + tensor var_2347_cast_fp16 = transpose(perm = var_2347_perm_0, x = var_2346_cast_fp16)[name = tensor("transpose_26")]; + tensor query_states_51_cast_fp16 = reshape(shape = var_2349, x = var_2347_cast_fp16)[name = tensor("query_states_51_cast_fp16")]; + tensor var_2351 = const()[name = tensor("op_2351"), val = tensor([20, -1, 64])]; + tensor var_2331_cast_fp16 = transpose(perm = var_2331_perm_0, x = var_2330_cast_fp16)[name = tensor("transpose_28")]; + tensor key_states_103_cast_fp16 = reshape(shape = var_2351, x = var_2331_cast_fp16)[name = tensor("key_states_103_cast_fp16")]; + tensor var_2353 = const()[name = tensor("op_2353"), val = tensor([20, -1, 64])]; + tensor var_2338_cast_fp16 = transpose(perm = var_2338_perm_0, x = var_2337_cast_fp16)[name = tensor("transpose_27")]; + tensor value_states_103_cast_fp16 = reshape(shape = var_2353, x = var_2338_cast_fp16)[name = tensor("value_states_103_cast_fp16")]; + tensor attn_weights_151_transpose_x_1 = const()[name = tensor("attn_weights_151_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_151_transpose_y_1 = const()[name = tensor("attn_weights_151_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_151_cast_fp16 = matmul(transpose_x = attn_weights_151_transpose_x_1, transpose_y = attn_weights_151_transpose_y_1, x = query_states_51_cast_fp16, y = key_states_103_cast_fp16)[name = tensor("attn_weights_151_cast_fp16")]; + tensor var_2358 = const()[name = tensor("op_2358"), val = tensor([1, 20, 77, 77])]; + tensor var_2359_cast_fp16 = reshape(shape = var_2358, x = attn_weights_151_cast_fp16)[name = tensor("op_2359_cast_fp16")]; + tensor attn_weights_153_cast_fp16 = add(x = var_2359_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_153_cast_fp16")]; + tensor var_2364 = const()[name = tensor("op_2364"), val = tensor([20, 77, 77])]; + tensor input_405_cast_fp16 = reshape(shape = var_2364, x = attn_weights_153_cast_fp16)[name = tensor("input_405_cast_fp16")]; + tensor input_407_cast_fp16 = softmax(axis = var_5, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; + tensor attn_output_151_transpose_x_0 = const()[name = tensor("attn_output_151_transpose_x_0"), val = tensor(false)]; + tensor attn_output_151_transpose_y_0 = const()[name = tensor("attn_output_151_transpose_y_0"), val = tensor(false)]; + tensor attn_output_151_cast_fp16 = matmul(transpose_x = attn_output_151_transpose_x_0, transpose_y = attn_output_151_transpose_y_0, x = input_407_cast_fp16, y = value_states_103_cast_fp16)[name = tensor("attn_output_151_cast_fp16")]; + tensor var_2369 = const()[name = tensor("op_2369"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_153_cast_fp16 = reshape(shape = var_2369, x = attn_output_151_cast_fp16)[name = tensor("attn_output_153_cast_fp16")]; + tensor attn_output_155_perm_0 = const()[name = tensor("attn_output_155_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2372 = const()[name = tensor("op_2372"), val = tensor([1, 77, 1280])]; + tensor attn_output_155_cast_fp16 = transpose(perm = attn_output_155_perm_0, x = attn_output_153_cast_fp16)[name = tensor("transpose_25")]; + tensor input_409_cast_fp16 = reshape(shape = var_2372, x = attn_output_155_cast_fp16)[name = tensor("input_409_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1120435072)))]; + tensor text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1123711936)))]; + tensor linear_153_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_self_attn_out_proj_weight_to_fp16, x = input_409_cast_fp16)[name = tensor("linear_153_cast_fp16")]; + tensor input_411_cast_fp16 = add(x = input_403_cast_fp16, y = linear_153_cast_fp16)[name = tensor("input_411_cast_fp16")]; + tensor input_413_axes_0 = const()[name = tensor("input_413_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1123714560)))]; + tensor text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1123717184)))]; + tensor input_413_cast_fp16 = layer_norm(axes = input_413_axes_0, beta = text_encoder_text_model_encoder_layers_25_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_25_layer_norm2_weight_to_fp16, x = input_411_cast_fp16)[name = tensor("input_413_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1123719808)))]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136827072)))]; + tensor linear_154_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_mlp_fc1_weight_to_fp16, x = input_413_cast_fp16)[name = tensor("linear_154_cast_fp16")]; + tensor input_417_mode_0 = const()[name = tensor("input_417_mode_0"), val = tensor("EXACT")]; + tensor input_417_cast_fp16 = gelu(mode = input_417_mode_0, x = linear_154_cast_fp16)[name = tensor("input_417_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1136837376)))]; + tensor text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1149944640)))]; + tensor linear_155_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_25_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_25_mlp_fc2_weight_to_fp16, x = input_417_cast_fp16)[name = tensor("linear_155_cast_fp16")]; + tensor input_419_cast_fp16 = add(x = input_411_cast_fp16, y = linear_155_cast_fp16)[name = tensor("input_419_cast_fp16")]; + tensor hidden_states_157_axes_0 = const()[name = tensor("hidden_states_157_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1149947264)))]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1149949888)))]; + tensor hidden_states_157_cast_fp16 = layer_norm(axes = hidden_states_157_axes_0, beta = text_encoder_text_model_encoder_layers_26_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_26_layer_norm1_weight_to_fp16, x = input_419_cast_fp16)[name = tensor("hidden_states_157_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1149952512)))]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1153229376)))]; + tensor linear_156_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_q_proj_weight_to_fp16, x = hidden_states_157_cast_fp16)[name = tensor("linear_156_cast_fp16")]; + tensor var_2411_to_fp16 = const()[name = tensor("op_2411_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_161_cast_fp16 = mul(x = linear_156_cast_fp16, y = var_2411_to_fp16)[name = tensor("tensor_161_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1153232000)))]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1156508864)))]; + tensor linear_157_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_k_proj_weight_to_fp16, x = hidden_states_157_cast_fp16)[name = tensor("linear_157_cast_fp16")]; + tensor var_2416 = const()[name = tensor("op_2416"), val = tensor([1, -1, 20, 64])]; + tensor var_2417_cast_fp16 = reshape(shape = var_2416, x = linear_157_cast_fp16)[name = tensor("op_2417_cast_fp16")]; + tensor var_2418_perm_0 = const()[name = tensor("op_2418_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1156511488)))]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1159788352)))]; + tensor linear_158_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_v_proj_weight_to_fp16, x = hidden_states_157_cast_fp16)[name = tensor("linear_158_cast_fp16")]; + tensor var_2423 = const()[name = tensor("op_2423"), val = tensor([1, -1, 20, 64])]; + tensor var_2424_cast_fp16 = reshape(shape = var_2423, x = linear_158_cast_fp16)[name = tensor("op_2424_cast_fp16")]; + tensor var_2425_perm_0 = const()[name = tensor("op_2425_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2432 = const()[name = tensor("op_2432"), val = tensor([1, 77, 20, 64])]; + tensor var_2433_cast_fp16 = reshape(shape = var_2432, x = tensor_161_cast_fp16)[name = tensor("op_2433_cast_fp16")]; + tensor var_2434_perm_0 = const()[name = tensor("op_2434_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2436 = const()[name = tensor("op_2436"), val = tensor([20, -1, 64])]; + tensor var_2434_cast_fp16 = transpose(perm = var_2434_perm_0, x = var_2433_cast_fp16)[name = tensor("transpose_22")]; + tensor query_states_53_cast_fp16 = reshape(shape = var_2436, x = var_2434_cast_fp16)[name = tensor("query_states_53_cast_fp16")]; + tensor var_2438 = const()[name = tensor("op_2438"), val = tensor([20, -1, 64])]; + tensor var_2418_cast_fp16 = transpose(perm = var_2418_perm_0, x = var_2417_cast_fp16)[name = tensor("transpose_24")]; + tensor key_states_107_cast_fp16 = reshape(shape = var_2438, x = var_2418_cast_fp16)[name = tensor("key_states_107_cast_fp16")]; + tensor var_2440 = const()[name = tensor("op_2440"), val = tensor([20, -1, 64])]; + tensor var_2425_cast_fp16 = transpose(perm = var_2425_perm_0, x = var_2424_cast_fp16)[name = tensor("transpose_23")]; + tensor value_states_107_cast_fp16 = reshape(shape = var_2440, x = var_2425_cast_fp16)[name = tensor("value_states_107_cast_fp16")]; + tensor attn_weights_157_transpose_x_1 = const()[name = tensor("attn_weights_157_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_157_transpose_y_1 = const()[name = tensor("attn_weights_157_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_157_cast_fp16 = matmul(transpose_x = attn_weights_157_transpose_x_1, transpose_y = attn_weights_157_transpose_y_1, x = query_states_53_cast_fp16, y = key_states_107_cast_fp16)[name = tensor("attn_weights_157_cast_fp16")]; + tensor var_2445 = const()[name = tensor("op_2445"), val = tensor([1, 20, 77, 77])]; + tensor var_2446_cast_fp16 = reshape(shape = var_2445, x = attn_weights_157_cast_fp16)[name = tensor("op_2446_cast_fp16")]; + tensor attn_weights_159_cast_fp16 = add(x = var_2446_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_159_cast_fp16")]; + tensor var_2451 = const()[name = tensor("op_2451"), val = tensor([20, 77, 77])]; + tensor input_421_cast_fp16 = reshape(shape = var_2451, x = attn_weights_159_cast_fp16)[name = tensor("input_421_cast_fp16")]; + tensor input_423_cast_fp16 = softmax(axis = var_5, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; + tensor attn_output_157_transpose_x_0 = const()[name = tensor("attn_output_157_transpose_x_0"), val = tensor(false)]; + tensor attn_output_157_transpose_y_0 = const()[name = tensor("attn_output_157_transpose_y_0"), val = tensor(false)]; + tensor attn_output_157_cast_fp16 = matmul(transpose_x = attn_output_157_transpose_x_0, transpose_y = attn_output_157_transpose_y_0, x = input_423_cast_fp16, y = value_states_107_cast_fp16)[name = tensor("attn_output_157_cast_fp16")]; + tensor var_2456 = const()[name = tensor("op_2456"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_159_cast_fp16 = reshape(shape = var_2456, x = attn_output_157_cast_fp16)[name = tensor("attn_output_159_cast_fp16")]; + tensor attn_output_161_perm_0 = const()[name = tensor("attn_output_161_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2459 = const()[name = tensor("op_2459"), val = tensor([1, 77, 1280])]; + tensor attn_output_161_cast_fp16 = transpose(perm = attn_output_161_perm_0, x = attn_output_159_cast_fp16)[name = tensor("transpose_21")]; + tensor input_425_cast_fp16 = reshape(shape = var_2459, x = attn_output_161_cast_fp16)[name = tensor("input_425_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1159790976)))]; + tensor text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1163067840)))]; + tensor linear_159_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_self_attn_out_proj_weight_to_fp16, x = input_425_cast_fp16)[name = tensor("linear_159_cast_fp16")]; + tensor input_427_cast_fp16 = add(x = input_419_cast_fp16, y = linear_159_cast_fp16)[name = tensor("input_427_cast_fp16")]; + tensor input_429_axes_0 = const()[name = tensor("input_429_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1163070464)))]; + tensor text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1163073088)))]; + tensor input_429_cast_fp16 = layer_norm(axes = input_429_axes_0, beta = text_encoder_text_model_encoder_layers_26_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_26_layer_norm2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("input_429_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1163075712)))]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1176182976)))]; + tensor linear_160_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_mlp_fc1_weight_to_fp16, x = input_429_cast_fp16)[name = tensor("linear_160_cast_fp16")]; + tensor input_433_mode_0 = const()[name = tensor("input_433_mode_0"), val = tensor("EXACT")]; + tensor input_433_cast_fp16 = gelu(mode = input_433_mode_0, x = linear_160_cast_fp16)[name = tensor("input_433_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1176193280)))]; + tensor text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1189300544)))]; + tensor linear_161_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_26_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_26_mlp_fc2_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("linear_161_cast_fp16")]; + tensor input_435_cast_fp16 = add(x = input_427_cast_fp16, y = linear_161_cast_fp16)[name = tensor("input_435_cast_fp16")]; + tensor hidden_states_163_axes_0 = const()[name = tensor("hidden_states_163_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1189303168)))]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1189305792)))]; + tensor hidden_states_163_cast_fp16 = layer_norm(axes = hidden_states_163_axes_0, beta = text_encoder_text_model_encoder_layers_27_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_27_layer_norm1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("hidden_states_163_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1189308416)))]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1192585280)))]; + tensor linear_162_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_q_proj_weight_to_fp16, x = hidden_states_163_cast_fp16)[name = tensor("linear_162_cast_fp16")]; + tensor var_2498_to_fp16 = const()[name = tensor("op_2498_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_167_cast_fp16 = mul(x = linear_162_cast_fp16, y = var_2498_to_fp16)[name = tensor("tensor_167_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1192587904)))]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1195864768)))]; + tensor linear_163_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_k_proj_weight_to_fp16, x = hidden_states_163_cast_fp16)[name = tensor("linear_163_cast_fp16")]; + tensor var_2503 = const()[name = tensor("op_2503"), val = tensor([1, -1, 20, 64])]; + tensor var_2504_cast_fp16 = reshape(shape = var_2503, x = linear_163_cast_fp16)[name = tensor("op_2504_cast_fp16")]; + tensor var_2505_perm_0 = const()[name = tensor("op_2505_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1195867392)))]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1199144256)))]; + tensor linear_164_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_v_proj_weight_to_fp16, x = hidden_states_163_cast_fp16)[name = tensor("linear_164_cast_fp16")]; + tensor var_2510 = const()[name = tensor("op_2510"), val = tensor([1, -1, 20, 64])]; + tensor var_2511_cast_fp16 = reshape(shape = var_2510, x = linear_164_cast_fp16)[name = tensor("op_2511_cast_fp16")]; + tensor var_2512_perm_0 = const()[name = tensor("op_2512_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2519 = const()[name = tensor("op_2519"), val = tensor([1, 77, 20, 64])]; + tensor var_2520_cast_fp16 = reshape(shape = var_2519, x = tensor_167_cast_fp16)[name = tensor("op_2520_cast_fp16")]; + tensor var_2521_perm_0 = const()[name = tensor("op_2521_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2523 = const()[name = tensor("op_2523"), val = tensor([20, -1, 64])]; + tensor var_2521_cast_fp16 = transpose(perm = var_2521_perm_0, x = var_2520_cast_fp16)[name = tensor("transpose_18")]; + tensor query_states_55_cast_fp16 = reshape(shape = var_2523, x = var_2521_cast_fp16)[name = tensor("query_states_55_cast_fp16")]; + tensor var_2525 = const()[name = tensor("op_2525"), val = tensor([20, -1, 64])]; + tensor var_2505_cast_fp16 = transpose(perm = var_2505_perm_0, x = var_2504_cast_fp16)[name = tensor("transpose_20")]; + tensor key_states_111_cast_fp16 = reshape(shape = var_2525, x = var_2505_cast_fp16)[name = tensor("key_states_111_cast_fp16")]; + tensor var_2527 = const()[name = tensor("op_2527"), val = tensor([20, -1, 64])]; + tensor var_2512_cast_fp16 = transpose(perm = var_2512_perm_0, x = var_2511_cast_fp16)[name = tensor("transpose_19")]; + tensor value_states_111_cast_fp16 = reshape(shape = var_2527, x = var_2512_cast_fp16)[name = tensor("value_states_111_cast_fp16")]; + tensor attn_weights_163_transpose_x_1 = const()[name = tensor("attn_weights_163_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_163_transpose_y_1 = const()[name = tensor("attn_weights_163_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_163_cast_fp16 = matmul(transpose_x = attn_weights_163_transpose_x_1, transpose_y = attn_weights_163_transpose_y_1, x = query_states_55_cast_fp16, y = key_states_111_cast_fp16)[name = tensor("attn_weights_163_cast_fp16")]; + tensor var_2532 = const()[name = tensor("op_2532"), val = tensor([1, 20, 77, 77])]; + tensor var_2533_cast_fp16 = reshape(shape = var_2532, x = attn_weights_163_cast_fp16)[name = tensor("op_2533_cast_fp16")]; + tensor attn_weights_165_cast_fp16 = add(x = var_2533_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_165_cast_fp16")]; + tensor var_2538 = const()[name = tensor("op_2538"), val = tensor([20, 77, 77])]; + tensor input_437_cast_fp16 = reshape(shape = var_2538, x = attn_weights_165_cast_fp16)[name = tensor("input_437_cast_fp16")]; + tensor input_439_cast_fp16 = softmax(axis = var_5, x = input_437_cast_fp16)[name = tensor("input_439_cast_fp16")]; + tensor attn_output_163_transpose_x_0 = const()[name = tensor("attn_output_163_transpose_x_0"), val = tensor(false)]; + tensor attn_output_163_transpose_y_0 = const()[name = tensor("attn_output_163_transpose_y_0"), val = tensor(false)]; + tensor attn_output_163_cast_fp16 = matmul(transpose_x = attn_output_163_transpose_x_0, transpose_y = attn_output_163_transpose_y_0, x = input_439_cast_fp16, y = value_states_111_cast_fp16)[name = tensor("attn_output_163_cast_fp16")]; + tensor var_2543 = const()[name = tensor("op_2543"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_165_cast_fp16 = reshape(shape = var_2543, x = attn_output_163_cast_fp16)[name = tensor("attn_output_165_cast_fp16")]; + tensor attn_output_167_perm_0 = const()[name = tensor("attn_output_167_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2546 = const()[name = tensor("op_2546"), val = tensor([1, 77, 1280])]; + tensor attn_output_167_cast_fp16 = transpose(perm = attn_output_167_perm_0, x = attn_output_165_cast_fp16)[name = tensor("transpose_17")]; + tensor input_441_cast_fp16 = reshape(shape = var_2546, x = attn_output_167_cast_fp16)[name = tensor("input_441_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1199146880)))]; + tensor text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1202423744)))]; + tensor linear_165_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_self_attn_out_proj_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("linear_165_cast_fp16")]; + tensor input_443_cast_fp16 = add(x = input_435_cast_fp16, y = linear_165_cast_fp16)[name = tensor("input_443_cast_fp16")]; + tensor input_445_axes_0 = const()[name = tensor("input_445_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1202426368)))]; + tensor text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1202428992)))]; + tensor input_445_cast_fp16 = layer_norm(axes = input_445_axes_0, beta = text_encoder_text_model_encoder_layers_27_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_27_layer_norm2_weight_to_fp16, x = input_443_cast_fp16)[name = tensor("input_445_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1202431616)))]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1215538880)))]; + tensor linear_166_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_mlp_fc1_weight_to_fp16, x = input_445_cast_fp16)[name = tensor("linear_166_cast_fp16")]; + tensor input_449_mode_0 = const()[name = tensor("input_449_mode_0"), val = tensor("EXACT")]; + tensor input_449_cast_fp16 = gelu(mode = input_449_mode_0, x = linear_166_cast_fp16)[name = tensor("input_449_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1215549184)))]; + tensor text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1228656448)))]; + tensor linear_167_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_27_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_27_mlp_fc2_weight_to_fp16, x = input_449_cast_fp16)[name = tensor("linear_167_cast_fp16")]; + tensor input_451_cast_fp16 = add(x = input_443_cast_fp16, y = linear_167_cast_fp16)[name = tensor("input_451_cast_fp16")]; + tensor hidden_states_169_axes_0 = const()[name = tensor("hidden_states_169_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1228659072)))]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1228661696)))]; + tensor hidden_states_169_cast_fp16 = layer_norm(axes = hidden_states_169_axes_0, beta = text_encoder_text_model_encoder_layers_28_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_28_layer_norm1_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("hidden_states_169_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1228664320)))]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1231941184)))]; + tensor linear_168_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_q_proj_weight_to_fp16, x = hidden_states_169_cast_fp16)[name = tensor("linear_168_cast_fp16")]; + tensor var_2585_to_fp16 = const()[name = tensor("op_2585_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_173_cast_fp16 = mul(x = linear_168_cast_fp16, y = var_2585_to_fp16)[name = tensor("tensor_173_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1231943808)))]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1235220672)))]; + tensor linear_169_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_k_proj_weight_to_fp16, x = hidden_states_169_cast_fp16)[name = tensor("linear_169_cast_fp16")]; + tensor var_2590 = const()[name = tensor("op_2590"), val = tensor([1, -1, 20, 64])]; + tensor var_2591_cast_fp16 = reshape(shape = var_2590, x = linear_169_cast_fp16)[name = tensor("op_2591_cast_fp16")]; + tensor var_2592_perm_0 = const()[name = tensor("op_2592_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1235223296)))]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1238500160)))]; + tensor linear_170_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_v_proj_weight_to_fp16, x = hidden_states_169_cast_fp16)[name = tensor("linear_170_cast_fp16")]; + tensor var_2597 = const()[name = tensor("op_2597"), val = tensor([1, -1, 20, 64])]; + tensor var_2598_cast_fp16 = reshape(shape = var_2597, x = linear_170_cast_fp16)[name = tensor("op_2598_cast_fp16")]; + tensor var_2599_perm_0 = const()[name = tensor("op_2599_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2606 = const()[name = tensor("op_2606"), val = tensor([1, 77, 20, 64])]; + tensor var_2607_cast_fp16 = reshape(shape = var_2606, x = tensor_173_cast_fp16)[name = tensor("op_2607_cast_fp16")]; + tensor var_2608_perm_0 = const()[name = tensor("op_2608_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2610 = const()[name = tensor("op_2610"), val = tensor([20, -1, 64])]; + tensor var_2608_cast_fp16 = transpose(perm = var_2608_perm_0, x = var_2607_cast_fp16)[name = tensor("transpose_14")]; + tensor query_states_57_cast_fp16 = reshape(shape = var_2610, x = var_2608_cast_fp16)[name = tensor("query_states_57_cast_fp16")]; + tensor var_2612 = const()[name = tensor("op_2612"), val = tensor([20, -1, 64])]; + tensor var_2592_cast_fp16 = transpose(perm = var_2592_perm_0, x = var_2591_cast_fp16)[name = tensor("transpose_16")]; + tensor key_states_115_cast_fp16 = reshape(shape = var_2612, x = var_2592_cast_fp16)[name = tensor("key_states_115_cast_fp16")]; + tensor var_2614 = const()[name = tensor("op_2614"), val = tensor([20, -1, 64])]; + tensor var_2599_cast_fp16 = transpose(perm = var_2599_perm_0, x = var_2598_cast_fp16)[name = tensor("transpose_15")]; + tensor value_states_115_cast_fp16 = reshape(shape = var_2614, x = var_2599_cast_fp16)[name = tensor("value_states_115_cast_fp16")]; + tensor attn_weights_169_transpose_x_1 = const()[name = tensor("attn_weights_169_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_169_transpose_y_1 = const()[name = tensor("attn_weights_169_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_169_cast_fp16 = matmul(transpose_x = attn_weights_169_transpose_x_1, transpose_y = attn_weights_169_transpose_y_1, x = query_states_57_cast_fp16, y = key_states_115_cast_fp16)[name = tensor("attn_weights_169_cast_fp16")]; + tensor var_2619 = const()[name = tensor("op_2619"), val = tensor([1, 20, 77, 77])]; + tensor var_2620_cast_fp16 = reshape(shape = var_2619, x = attn_weights_169_cast_fp16)[name = tensor("op_2620_cast_fp16")]; + tensor attn_weights_171_cast_fp16 = add(x = var_2620_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_171_cast_fp16")]; + tensor var_2625 = const()[name = tensor("op_2625"), val = tensor([20, 77, 77])]; + tensor input_453_cast_fp16 = reshape(shape = var_2625, x = attn_weights_171_cast_fp16)[name = tensor("input_453_cast_fp16")]; + tensor input_455_cast_fp16 = softmax(axis = var_5, x = input_453_cast_fp16)[name = tensor("input_455_cast_fp16")]; + tensor attn_output_169_transpose_x_0 = const()[name = tensor("attn_output_169_transpose_x_0"), val = tensor(false)]; + tensor attn_output_169_transpose_y_0 = const()[name = tensor("attn_output_169_transpose_y_0"), val = tensor(false)]; + tensor attn_output_169_cast_fp16 = matmul(transpose_x = attn_output_169_transpose_x_0, transpose_y = attn_output_169_transpose_y_0, x = input_455_cast_fp16, y = value_states_115_cast_fp16)[name = tensor("attn_output_169_cast_fp16")]; + tensor var_2630 = const()[name = tensor("op_2630"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_171_cast_fp16 = reshape(shape = var_2630, x = attn_output_169_cast_fp16)[name = tensor("attn_output_171_cast_fp16")]; + tensor attn_output_173_perm_0 = const()[name = tensor("attn_output_173_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2633 = const()[name = tensor("op_2633"), val = tensor([1, 77, 1280])]; + tensor attn_output_173_cast_fp16 = transpose(perm = attn_output_173_perm_0, x = attn_output_171_cast_fp16)[name = tensor("transpose_13")]; + tensor input_457_cast_fp16 = reshape(shape = var_2633, x = attn_output_173_cast_fp16)[name = tensor("input_457_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1238502784)))]; + tensor text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1241779648)))]; + tensor linear_171_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_self_attn_out_proj_weight_to_fp16, x = input_457_cast_fp16)[name = tensor("linear_171_cast_fp16")]; + tensor input_459_cast_fp16 = add(x = input_451_cast_fp16, y = linear_171_cast_fp16)[name = tensor("input_459_cast_fp16")]; + tensor input_461_axes_0 = const()[name = tensor("input_461_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1241782272)))]; + tensor text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1241784896)))]; + tensor input_461_cast_fp16 = layer_norm(axes = input_461_axes_0, beta = text_encoder_text_model_encoder_layers_28_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_28_layer_norm2_weight_to_fp16, x = input_459_cast_fp16)[name = tensor("input_461_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1241787520)))]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1254894784)))]; + tensor linear_172_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_mlp_fc1_weight_to_fp16, x = input_461_cast_fp16)[name = tensor("linear_172_cast_fp16")]; + tensor input_465_mode_0 = const()[name = tensor("input_465_mode_0"), val = tensor("EXACT")]; + tensor input_465_cast_fp16 = gelu(mode = input_465_mode_0, x = linear_172_cast_fp16)[name = tensor("input_465_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1254905088)))]; + tensor text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1268012352)))]; + tensor linear_173_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_28_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_28_mlp_fc2_weight_to_fp16, x = input_465_cast_fp16)[name = tensor("linear_173_cast_fp16")]; + tensor input_467_cast_fp16 = add(x = input_459_cast_fp16, y = linear_173_cast_fp16)[name = tensor("input_467_cast_fp16")]; + tensor hidden_states_175_axes_0 = const()[name = tensor("hidden_states_175_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1268014976)))]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1268017600)))]; + tensor hidden_states_175_cast_fp16 = layer_norm(axes = hidden_states_175_axes_0, beta = text_encoder_text_model_encoder_layers_29_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_29_layer_norm1_weight_to_fp16, x = input_467_cast_fp16)[name = tensor("hidden_states_175_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1268020224)))]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1271297088)))]; + tensor linear_174_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_q_proj_weight_to_fp16, x = hidden_states_175_cast_fp16)[name = tensor("linear_174_cast_fp16")]; + tensor var_2672_to_fp16 = const()[name = tensor("op_2672_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_179_cast_fp16 = mul(x = linear_174_cast_fp16, y = var_2672_to_fp16)[name = tensor("tensor_179_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1271299712)))]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1274576576)))]; + tensor linear_175_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_k_proj_weight_to_fp16, x = hidden_states_175_cast_fp16)[name = tensor("linear_175_cast_fp16")]; + tensor var_2677 = const()[name = tensor("op_2677"), val = tensor([1, -1, 20, 64])]; + tensor var_2678_cast_fp16 = reshape(shape = var_2677, x = linear_175_cast_fp16)[name = tensor("op_2678_cast_fp16")]; + tensor var_2679_perm_0 = const()[name = tensor("op_2679_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1274579200)))]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1277856064)))]; + tensor linear_176_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_v_proj_weight_to_fp16, x = hidden_states_175_cast_fp16)[name = tensor("linear_176_cast_fp16")]; + tensor var_2684 = const()[name = tensor("op_2684"), val = tensor([1, -1, 20, 64])]; + tensor var_2685_cast_fp16 = reshape(shape = var_2684, x = linear_176_cast_fp16)[name = tensor("op_2685_cast_fp16")]; + tensor var_2686_perm_0 = const()[name = tensor("op_2686_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2693 = const()[name = tensor("op_2693"), val = tensor([1, 77, 20, 64])]; + tensor var_2694_cast_fp16 = reshape(shape = var_2693, x = tensor_179_cast_fp16)[name = tensor("op_2694_cast_fp16")]; + tensor var_2695_perm_0 = const()[name = tensor("op_2695_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2697 = const()[name = tensor("op_2697"), val = tensor([20, -1, 64])]; + tensor var_2695_cast_fp16 = transpose(perm = var_2695_perm_0, x = var_2694_cast_fp16)[name = tensor("transpose_10")]; + tensor query_states_59_cast_fp16 = reshape(shape = var_2697, x = var_2695_cast_fp16)[name = tensor("query_states_59_cast_fp16")]; + tensor var_2699 = const()[name = tensor("op_2699"), val = tensor([20, -1, 64])]; + tensor var_2679_cast_fp16 = transpose(perm = var_2679_perm_0, x = var_2678_cast_fp16)[name = tensor("transpose_12")]; + tensor key_states_119_cast_fp16 = reshape(shape = var_2699, x = var_2679_cast_fp16)[name = tensor("key_states_119_cast_fp16")]; + tensor var_2701 = const()[name = tensor("op_2701"), val = tensor([20, -1, 64])]; + tensor var_2686_cast_fp16 = transpose(perm = var_2686_perm_0, x = var_2685_cast_fp16)[name = tensor("transpose_11")]; + tensor value_states_119_cast_fp16 = reshape(shape = var_2701, x = var_2686_cast_fp16)[name = tensor("value_states_119_cast_fp16")]; + tensor attn_weights_175_transpose_x_1 = const()[name = tensor("attn_weights_175_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_175_transpose_y_1 = const()[name = tensor("attn_weights_175_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_175_cast_fp16 = matmul(transpose_x = attn_weights_175_transpose_x_1, transpose_y = attn_weights_175_transpose_y_1, x = query_states_59_cast_fp16, y = key_states_119_cast_fp16)[name = tensor("attn_weights_175_cast_fp16")]; + tensor var_2706 = const()[name = tensor("op_2706"), val = tensor([1, 20, 77, 77])]; + tensor var_2707_cast_fp16 = reshape(shape = var_2706, x = attn_weights_175_cast_fp16)[name = tensor("op_2707_cast_fp16")]; + tensor attn_weights_177_cast_fp16 = add(x = var_2707_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_177_cast_fp16")]; + tensor var_2712 = const()[name = tensor("op_2712"), val = tensor([20, 77, 77])]; + tensor input_469_cast_fp16 = reshape(shape = var_2712, x = attn_weights_177_cast_fp16)[name = tensor("input_469_cast_fp16")]; + tensor input_471_cast_fp16 = softmax(axis = var_5, x = input_469_cast_fp16)[name = tensor("input_471_cast_fp16")]; + tensor attn_output_175_transpose_x_0 = const()[name = tensor("attn_output_175_transpose_x_0"), val = tensor(false)]; + tensor attn_output_175_transpose_y_0 = const()[name = tensor("attn_output_175_transpose_y_0"), val = tensor(false)]; + tensor attn_output_175_cast_fp16 = matmul(transpose_x = attn_output_175_transpose_x_0, transpose_y = attn_output_175_transpose_y_0, x = input_471_cast_fp16, y = value_states_119_cast_fp16)[name = tensor("attn_output_175_cast_fp16")]; + tensor var_2717 = const()[name = tensor("op_2717"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_177_cast_fp16 = reshape(shape = var_2717, x = attn_output_175_cast_fp16)[name = tensor("attn_output_177_cast_fp16")]; + tensor attn_output_179_perm_0 = const()[name = tensor("attn_output_179_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2720 = const()[name = tensor("op_2720"), val = tensor([1, 77, 1280])]; + tensor attn_output_179_cast_fp16 = transpose(perm = attn_output_179_perm_0, x = attn_output_177_cast_fp16)[name = tensor("transpose_9")]; + tensor input_473_cast_fp16 = reshape(shape = var_2720, x = attn_output_179_cast_fp16)[name = tensor("input_473_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1277858688)))]; + tensor text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1281135552)))]; + tensor linear_177_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_self_attn_out_proj_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("linear_177_cast_fp16")]; + tensor input_475_cast_fp16 = add(x = input_467_cast_fp16, y = linear_177_cast_fp16)[name = tensor("input_475_cast_fp16")]; + tensor input_477_axes_0 = const()[name = tensor("input_477_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1281138176)))]; + tensor text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1281140800)))]; + tensor input_477_cast_fp16 = layer_norm(axes = input_477_axes_0, beta = text_encoder_text_model_encoder_layers_29_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_29_layer_norm2_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("input_477_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1281143424)))]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1294250688)))]; + tensor linear_178_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_mlp_fc1_weight_to_fp16, x = input_477_cast_fp16)[name = tensor("linear_178_cast_fp16")]; + tensor input_481_mode_0 = const()[name = tensor("input_481_mode_0"), val = tensor("EXACT")]; + tensor input_481_cast_fp16 = gelu(mode = input_481_mode_0, x = linear_178_cast_fp16)[name = tensor("input_481_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1294260992)))]; + tensor text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1307368256)))]; + tensor linear_179_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_29_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_29_mlp_fc2_weight_to_fp16, x = input_481_cast_fp16)[name = tensor("linear_179_cast_fp16")]; + tensor input_483_cast_fp16 = add(x = input_475_cast_fp16, y = linear_179_cast_fp16)[name = tensor("input_483_cast_fp16")]; + tensor hidden_states_181_axes_0 = const()[name = tensor("hidden_states_181_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1307370880)))]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1307373504)))]; + tensor hidden_states_181_cast_fp16 = layer_norm(axes = hidden_states_181_axes_0, beta = text_encoder_text_model_encoder_layers_30_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_30_layer_norm1_weight_to_fp16, x = input_483_cast_fp16)[name = tensor("hidden_states_181_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1307376128)))]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1310652992)))]; + tensor linear_180_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_q_proj_weight_to_fp16, x = hidden_states_181_cast_fp16)[name = tensor("linear_180_cast_fp16")]; + tensor var_2759_to_fp16 = const()[name = tensor("op_2759_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_185_cast_fp16 = mul(x = linear_180_cast_fp16, y = var_2759_to_fp16)[name = tensor("tensor_185_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1310655616)))]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1313932480)))]; + tensor linear_181_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_k_proj_weight_to_fp16, x = hidden_states_181_cast_fp16)[name = tensor("linear_181_cast_fp16")]; + tensor var_2764 = const()[name = tensor("op_2764"), val = tensor([1, -1, 20, 64])]; + tensor var_2765_cast_fp16 = reshape(shape = var_2764, x = linear_181_cast_fp16)[name = tensor("op_2765_cast_fp16")]; + tensor var_2766_perm_0 = const()[name = tensor("op_2766_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1313935104)))]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1317211968)))]; + tensor linear_182_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_v_proj_weight_to_fp16, x = hidden_states_181_cast_fp16)[name = tensor("linear_182_cast_fp16")]; + tensor var_2771 = const()[name = tensor("op_2771"), val = tensor([1, -1, 20, 64])]; + tensor var_2772_cast_fp16 = reshape(shape = var_2771, x = linear_182_cast_fp16)[name = tensor("op_2772_cast_fp16")]; + tensor var_2773_perm_0 = const()[name = tensor("op_2773_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2780 = const()[name = tensor("op_2780"), val = tensor([1, 77, 20, 64])]; + tensor var_2781_cast_fp16 = reshape(shape = var_2780, x = tensor_185_cast_fp16)[name = tensor("op_2781_cast_fp16")]; + tensor var_2782_perm_0 = const()[name = tensor("op_2782_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2784 = const()[name = tensor("op_2784"), val = tensor([20, -1, 64])]; + tensor var_2782_cast_fp16 = transpose(perm = var_2782_perm_0, x = var_2781_cast_fp16)[name = tensor("transpose_6")]; + tensor query_states_61_cast_fp16 = reshape(shape = var_2784, x = var_2782_cast_fp16)[name = tensor("query_states_61_cast_fp16")]; + tensor var_2786 = const()[name = tensor("op_2786"), val = tensor([20, -1, 64])]; + tensor var_2766_cast_fp16 = transpose(perm = var_2766_perm_0, x = var_2765_cast_fp16)[name = tensor("transpose_8")]; + tensor key_states_123_cast_fp16 = reshape(shape = var_2786, x = var_2766_cast_fp16)[name = tensor("key_states_123_cast_fp16")]; + tensor var_2788 = const()[name = tensor("op_2788"), val = tensor([20, -1, 64])]; + tensor var_2773_cast_fp16 = transpose(perm = var_2773_perm_0, x = var_2772_cast_fp16)[name = tensor("transpose_7")]; + tensor value_states_123_cast_fp16 = reshape(shape = var_2788, x = var_2773_cast_fp16)[name = tensor("value_states_123_cast_fp16")]; + tensor attn_weights_181_transpose_x_1 = const()[name = tensor("attn_weights_181_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_181_transpose_y_1 = const()[name = tensor("attn_weights_181_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_181_cast_fp16 = matmul(transpose_x = attn_weights_181_transpose_x_1, transpose_y = attn_weights_181_transpose_y_1, x = query_states_61_cast_fp16, y = key_states_123_cast_fp16)[name = tensor("attn_weights_181_cast_fp16")]; + tensor var_2793 = const()[name = tensor("op_2793"), val = tensor([1, 20, 77, 77])]; + tensor var_2794_cast_fp16 = reshape(shape = var_2793, x = attn_weights_181_cast_fp16)[name = tensor("op_2794_cast_fp16")]; + tensor attn_weights_183_cast_fp16 = add(x = var_2794_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_183_cast_fp16")]; + tensor var_2799 = const()[name = tensor("op_2799"), val = tensor([20, 77, 77])]; + tensor input_485_cast_fp16 = reshape(shape = var_2799, x = attn_weights_183_cast_fp16)[name = tensor("input_485_cast_fp16")]; + tensor input_487_cast_fp16 = softmax(axis = var_5, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; + tensor attn_output_181_transpose_x_0 = const()[name = tensor("attn_output_181_transpose_x_0"), val = tensor(false)]; + tensor attn_output_181_transpose_y_0 = const()[name = tensor("attn_output_181_transpose_y_0"), val = tensor(false)]; + tensor attn_output_181_cast_fp16 = matmul(transpose_x = attn_output_181_transpose_x_0, transpose_y = attn_output_181_transpose_y_0, x = input_487_cast_fp16, y = value_states_123_cast_fp16)[name = tensor("attn_output_181_cast_fp16")]; + tensor var_2804 = const()[name = tensor("op_2804"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_183_cast_fp16 = reshape(shape = var_2804, x = attn_output_181_cast_fp16)[name = tensor("attn_output_183_cast_fp16")]; + tensor attn_output_185_perm_0 = const()[name = tensor("attn_output_185_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2807 = const()[name = tensor("op_2807"), val = tensor([1, 77, 1280])]; + tensor attn_output_185_cast_fp16 = transpose(perm = attn_output_185_perm_0, x = attn_output_183_cast_fp16)[name = tensor("transpose_5")]; + tensor input_489_cast_fp16 = reshape(shape = var_2807, x = attn_output_185_cast_fp16)[name = tensor("input_489_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1317214592)))]; + tensor text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1320491456)))]; + tensor linear_183_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_self_attn_out_proj_weight_to_fp16, x = input_489_cast_fp16)[name = tensor("linear_183_cast_fp16")]; + tensor input_491_cast_fp16 = add(x = input_483_cast_fp16, y = linear_183_cast_fp16)[name = tensor("input_491_cast_fp16")]; + tensor input_493_axes_0 = const()[name = tensor("input_493_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1320494080)))]; + tensor text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1320496704)))]; + tensor input_493_cast_fp16 = layer_norm(axes = input_493_axes_0, beta = text_encoder_text_model_encoder_layers_30_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_30_layer_norm2_weight_to_fp16, x = input_491_cast_fp16)[name = tensor("input_493_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1320499328)))]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1333606592)))]; + tensor linear_184_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_mlp_fc1_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("linear_184_cast_fp16")]; + tensor input_497_mode_0 = const()[name = tensor("input_497_mode_0"), val = tensor("EXACT")]; + tensor input_497_cast_fp16 = gelu(mode = input_497_mode_0, x = linear_184_cast_fp16)[name = tensor("input_497_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1333616896)))]; + tensor text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1346724160)))]; + tensor linear_185_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_30_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_30_mlp_fc2_weight_to_fp16, x = input_497_cast_fp16)[name = tensor("linear_185_cast_fp16")]; + tensor input_499_cast_fp16 = add(x = input_491_cast_fp16, y = linear_185_cast_fp16)[name = tensor("input_499_cast_fp16")]; + tensor input_499_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("input_499_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor hidden_states_187_axes_0 = const()[name = tensor("hidden_states_187_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1346726784)))]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1346729408)))]; + tensor hidden_states_187_cast_fp16 = layer_norm(axes = hidden_states_187_axes_0, beta = text_encoder_text_model_encoder_layers_31_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_31_layer_norm1_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("hidden_states_187_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1346732032)))]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1350008896)))]; + tensor linear_186_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_q_proj_weight_to_fp16, x = hidden_states_187_cast_fp16)[name = tensor("linear_186_cast_fp16")]; + tensor var_2846_to_fp16 = const()[name = tensor("op_2846_to_fp16"), val = tensor(0x1p-3)]; + tensor tensor_cast_fp16 = mul(x = linear_186_cast_fp16, y = var_2846_to_fp16)[name = tensor("tensor_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1350011520)))]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1353288384)))]; + tensor linear_187_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_k_proj_weight_to_fp16, x = hidden_states_187_cast_fp16)[name = tensor("linear_187_cast_fp16")]; + tensor var_2851 = const()[name = tensor("op_2851"), val = tensor([1, -1, 20, 64])]; + tensor var_2852_cast_fp16 = reshape(shape = var_2851, x = linear_187_cast_fp16)[name = tensor("op_2852_cast_fp16")]; + tensor var_2853_perm_0 = const()[name = tensor("op_2853_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1353291008)))]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1356567872)))]; + tensor linear_188_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_v_proj_weight_to_fp16, x = hidden_states_187_cast_fp16)[name = tensor("linear_188_cast_fp16")]; + tensor var_2858 = const()[name = tensor("op_2858"), val = tensor([1, -1, 20, 64])]; + tensor var_2859_cast_fp16 = reshape(shape = var_2858, x = linear_188_cast_fp16)[name = tensor("op_2859_cast_fp16")]; + tensor var_2860_perm_0 = const()[name = tensor("op_2860_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2867 = const()[name = tensor("op_2867"), val = tensor([1, 77, 20, 64])]; + tensor var_2868_cast_fp16 = reshape(shape = var_2867, x = tensor_cast_fp16)[name = tensor("op_2868_cast_fp16")]; + tensor var_2869_perm_0 = const()[name = tensor("op_2869_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2871 = const()[name = tensor("op_2871"), val = tensor([20, -1, 64])]; + tensor var_2869_cast_fp16 = transpose(perm = var_2869_perm_0, x = var_2868_cast_fp16)[name = tensor("transpose_2")]; + tensor query_states_cast_fp16 = reshape(shape = var_2871, x = var_2869_cast_fp16)[name = tensor("query_states_cast_fp16")]; + tensor var_2873 = const()[name = tensor("op_2873"), val = tensor([20, -1, 64])]; + tensor var_2853_cast_fp16 = transpose(perm = var_2853_perm_0, x = var_2852_cast_fp16)[name = tensor("transpose_4")]; + tensor key_states_cast_fp16 = reshape(shape = var_2873, x = var_2853_cast_fp16)[name = tensor("key_states_cast_fp16")]; + tensor var_2875 = const()[name = tensor("op_2875"), val = tensor([20, -1, 64])]; + tensor var_2860_cast_fp16 = transpose(perm = var_2860_perm_0, x = var_2859_cast_fp16)[name = tensor("transpose_3")]; + tensor value_states_cast_fp16 = reshape(shape = var_2875, x = var_2860_cast_fp16)[name = tensor("value_states_cast_fp16")]; + tensor attn_weights_187_transpose_x_1 = const()[name = tensor("attn_weights_187_transpose_x_1"), val = tensor(false)]; + tensor attn_weights_187_transpose_y_1 = const()[name = tensor("attn_weights_187_transpose_y_1"), val = tensor(true)]; + tensor attn_weights_187_cast_fp16 = matmul(transpose_x = attn_weights_187_transpose_x_1, transpose_y = attn_weights_187_transpose_y_1, x = query_states_cast_fp16, y = key_states_cast_fp16)[name = tensor("attn_weights_187_cast_fp16")]; + tensor var_2880 = const()[name = tensor("op_2880"), val = tensor([1, 20, 77, 77])]; + tensor var_2881_cast_fp16 = reshape(shape = var_2880, x = attn_weights_187_cast_fp16)[name = tensor("op_2881_cast_fp16")]; + tensor attn_weights_189_cast_fp16 = add(x = var_2881_cast_fp16, y = var_58_to_fp16)[name = tensor("attn_weights_189_cast_fp16")]; + tensor var_2886 = const()[name = tensor("op_2886"), val = tensor([20, 77, 77])]; + tensor input_501_cast_fp16 = reshape(shape = var_2886, x = attn_weights_189_cast_fp16)[name = tensor("input_501_cast_fp16")]; + tensor input_503_cast_fp16 = softmax(axis = var_5, x = input_501_cast_fp16)[name = tensor("input_503_cast_fp16")]; + tensor attn_output_187_transpose_x_0 = const()[name = tensor("attn_output_187_transpose_x_0"), val = tensor(false)]; + tensor attn_output_187_transpose_y_0 = const()[name = tensor("attn_output_187_transpose_y_0"), val = tensor(false)]; + tensor attn_output_187_cast_fp16 = matmul(transpose_x = attn_output_187_transpose_x_0, transpose_y = attn_output_187_transpose_y_0, x = input_503_cast_fp16, y = value_states_cast_fp16)[name = tensor("attn_output_187_cast_fp16")]; + tensor var_2891 = const()[name = tensor("op_2891"), val = tensor([1, 20, 77, 64])]; + tensor attn_output_189_cast_fp16 = reshape(shape = var_2891, x = attn_output_187_cast_fp16)[name = tensor("attn_output_189_cast_fp16")]; + tensor attn_output_perm_0 = const()[name = tensor("attn_output_perm_0"), val = tensor([0, 2, 1, 3])]; + tensor var_2894 = const()[name = tensor("op_2894"), val = tensor([1, 77, 1280])]; + tensor attn_output_cast_fp16 = transpose(perm = attn_output_perm_0, x = attn_output_189_cast_fp16)[name = tensor("transpose_1")]; + tensor input_505_cast_fp16 = reshape(shape = var_2894, x = attn_output_cast_fp16)[name = tensor("input_505_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1356570496)))]; + tensor text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1359847360)))]; + tensor linear_189_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_self_attn_out_proj_weight_to_fp16, x = input_505_cast_fp16)[name = tensor("linear_189_cast_fp16")]; + tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_189_cast_fp16)[name = tensor("input_507_cast_fp16")]; + tensor input_509_axes_0 = const()[name = tensor("input_509_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1359849984)))]; + tensor text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1359852608)))]; + tensor input_509_cast_fp16 = layer_norm(axes = input_509_axes_0, beta = text_encoder_text_model_encoder_layers_31_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_31_layer_norm2_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("input_509_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1359855232)))]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1372962496)))]; + tensor linear_190_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_mlp_fc1_weight_to_fp16, x = input_509_cast_fp16)[name = tensor("linear_190_cast_fp16")]; + tensor input_513_mode_0 = const()[name = tensor("input_513_mode_0"), val = tensor("EXACT")]; + tensor input_513_cast_fp16 = gelu(mode = input_513_mode_0, x = linear_190_cast_fp16)[name = tensor("input_513_cast_fp16")]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1372972800)))]; + tensor text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1386080064)))]; + tensor linear_191_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_31_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_31_mlp_fc2_weight_to_fp16, x = input_513_cast_fp16)[name = tensor("linear_191_cast_fp16")]; + tensor input_515_cast_fp16 = add(x = input_507_cast_fp16, y = linear_191_cast_fp16)[name = tensor("input_515_cast_fp16")]; + tensor last_hidden_state_axes_0 = const()[name = tensor("last_hidden_state_axes_0"), val = tensor([-1])]; + tensor text_encoder_text_model_final_layer_norm_weight_to_fp16 = const()[name = tensor("text_encoder_text_model_final_layer_norm_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1386082688)))]; + tensor text_encoder_text_model_final_layer_norm_bias_to_fp16 = const()[name = tensor("text_encoder_text_model_final_layer_norm_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1386085312)))]; + tensor last_hidden_state_cast_fp16 = layer_norm(axes = last_hidden_state_axes_0, beta = text_encoder_text_model_final_layer_norm_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_final_layer_norm_weight_to_fp16, x = input_515_cast_fp16)[name = tensor("last_hidden_state_cast_fp16")]; + tensor var_2922 = const()[name = tensor("op_2922"), val = tensor([0])]; + tensor var_2924 = reduce_argmax(axis = var_5, keep_dims = var_6, x = cast_1)[name = tensor("op_2924")]; + tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(1)]; + tensor stack_0 = stack(axis = stack_0_axis_0, values = (var_2922, var_2924))[name = tensor("stack_0")]; + tensor input_transpose_batch_dims_0 = const()[name = tensor("input_transpose_batch_dims_0"), val = tensor(0)]; + tensor input_transpose_cast_fp16 = gather_nd(batch_dims = input_transpose_batch_dims_0, indices = stack_0, x = last_hidden_state_cast_fp16)[name = tensor("input_transpose_cast_fp16")]; + tensor text_encoder_text_projection_weight_to_fp16 = const()[name = tensor("text_encoder_text_projection_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1386087936)))]; + tensor linear_192_bias_0_to_fp16 = const()[name = tensor("linear_192_bias_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1389364800)))]; + tensor linear_192_cast_fp16 = linear(bias = linear_192_bias_0_to_fp16, weight = text_encoder_text_projection_weight_to_fp16, x = input_transpose_cast_fp16)[name = tensor("linear_192_cast_fp16")]; + tensor linear_192_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("linear_192_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; + tensor pooled_outputs = cast(dtype = linear_192_cast_fp16_to_fp32_dtype_0, x = linear_192_cast_fp16)[name = tensor("cast_327")]; + tensor hidden_embeds = cast(dtype = input_499_cast_fp16_to_fp32_dtype_0, x = input_499_cast_fp16)[name = tensor("cast_328")]; + } -> (hidden_embeds, pooled_outputs); +} \ No newline at end of file